var/home/core/zuul-output/0000755000175000017500000000000015126107461014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015126114070015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003501101315126114061017664 0ustar rootrootJan 03 03:41:03 crc systemd[1]: Starting Kubernetes Kubelet... Jan 03 03:41:03 crc restorecon[4680]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:03 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 03 03:41:04 crc restorecon[4680]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 03 03:41:04 crc kubenswrapper[4921]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.695353 4921 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701340 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701379 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701386 4921 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701393 4921 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701399 4921 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701405 4921 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701411 4921 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701417 4921 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701422 4921 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701428 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701433 4921 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701439 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701445 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701451 4921 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701456 4921 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701461 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701467 4921 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701472 4921 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701486 4921 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701492 4921 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701498 4921 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701505 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701510 4921 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701515 4921 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701520 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701526 4921 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701531 4921 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701536 4921 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701542 4921 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701549 4921 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701559 4921 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701568 4921 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701575 4921 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701583 4921 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701589 4921 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701596 4921 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701602 4921 feature_gate.go:330] unrecognized feature gate: Example Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701609 4921 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701617 4921 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701625 4921 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701632 4921 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701639 4921 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701645 4921 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701651 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701657 4921 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701663 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701668 4921 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701674 4921 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701680 4921 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701687 4921 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701693 4921 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701700 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701707 4921 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701715 4921 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701721 4921 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701727 4921 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701732 4921 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701738 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701743 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701748 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701753 4921 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701759 4921 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701764 4921 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701771 4921 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701777 4921 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701782 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701787 4921 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701793 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701798 4921 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701804 4921 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.701809 4921 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.701955 4921 flags.go:64] FLAG: --address="0.0.0.0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.701967 4921 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.701977 4921 flags.go:64] FLAG: --anonymous-auth="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.701987 4921 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.701995 4921 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702002 4921 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702011 4921 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702024 4921 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702032 4921 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702039 4921 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702047 4921 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702054 4921 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702060 4921 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702068 4921 flags.go:64] FLAG: --cgroup-root="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702077 4921 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702084 4921 flags.go:64] FLAG: --client-ca-file="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702092 4921 flags.go:64] FLAG: --cloud-config="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702098 4921 flags.go:64] FLAG: --cloud-provider="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702105 4921 flags.go:64] FLAG: --cluster-dns="[]" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702114 4921 flags.go:64] FLAG: --cluster-domain="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702120 4921 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702127 4921 flags.go:64] FLAG: --config-dir="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702133 4921 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702140 4921 flags.go:64] FLAG: --container-log-max-files="5" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702149 4921 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702156 4921 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702162 4921 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702170 4921 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702177 4921 flags.go:64] FLAG: --contention-profiling="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702184 4921 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702190 4921 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702198 4921 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702211 4921 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702219 4921 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702226 4921 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702232 4921 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702239 4921 flags.go:64] FLAG: --enable-load-reader="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702246 4921 flags.go:64] FLAG: --enable-server="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702252 4921 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702260 4921 flags.go:64] FLAG: --event-burst="100" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702286 4921 flags.go:64] FLAG: --event-qps="50" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702293 4921 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702299 4921 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702305 4921 flags.go:64] FLAG: --eviction-hard="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702330 4921 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702337 4921 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702344 4921 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702351 4921 flags.go:64] FLAG: --eviction-soft="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702358 4921 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702365 4921 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702372 4921 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702378 4921 flags.go:64] FLAG: --experimental-mounter-path="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702385 4921 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702391 4921 flags.go:64] FLAG: --fail-swap-on="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702397 4921 flags.go:64] FLAG: --feature-gates="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702405 4921 flags.go:64] FLAG: --file-check-frequency="20s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702411 4921 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702418 4921 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702425 4921 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702431 4921 flags.go:64] FLAG: --healthz-port="10248" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702438 4921 flags.go:64] FLAG: --help="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702444 4921 flags.go:64] FLAG: --hostname-override="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702451 4921 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702458 4921 flags.go:64] FLAG: --http-check-frequency="20s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702465 4921 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702472 4921 flags.go:64] FLAG: --image-credential-provider-config="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702478 4921 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702485 4921 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702502 4921 flags.go:64] FLAG: --image-service-endpoint="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702509 4921 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702516 4921 flags.go:64] FLAG: --kube-api-burst="100" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702522 4921 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702529 4921 flags.go:64] FLAG: --kube-api-qps="50" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702535 4921 flags.go:64] FLAG: --kube-reserved="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702541 4921 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702547 4921 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702554 4921 flags.go:64] FLAG: --kubelet-cgroups="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702560 4921 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702567 4921 flags.go:64] FLAG: --lock-file="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702573 4921 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702580 4921 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702588 4921 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702598 4921 flags.go:64] FLAG: --log-json-split-stream="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702604 4921 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702610 4921 flags.go:64] FLAG: --log-text-split-stream="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702616 4921 flags.go:64] FLAG: --logging-format="text" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702623 4921 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702629 4921 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702636 4921 flags.go:64] FLAG: --manifest-url="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702642 4921 flags.go:64] FLAG: --manifest-url-header="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702651 4921 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702657 4921 flags.go:64] FLAG: --max-open-files="1000000" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702680 4921 flags.go:64] FLAG: --max-pods="110" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702687 4921 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702693 4921 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702699 4921 flags.go:64] FLAG: --memory-manager-policy="None" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702705 4921 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702712 4921 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702718 4921 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702724 4921 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702740 4921 flags.go:64] FLAG: --node-status-max-images="50" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702747 4921 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702753 4921 flags.go:64] FLAG: --oom-score-adj="-999" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702759 4921 flags.go:64] FLAG: --pod-cidr="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702774 4921 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702787 4921 flags.go:64] FLAG: --pod-manifest-path="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702794 4921 flags.go:64] FLAG: --pod-max-pids="-1" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702800 4921 flags.go:64] FLAG: --pods-per-core="0" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702806 4921 flags.go:64] FLAG: --port="10250" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702812 4921 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702820 4921 flags.go:64] FLAG: --provider-id="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702828 4921 flags.go:64] FLAG: --qos-reserved="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702834 4921 flags.go:64] FLAG: --read-only-port="10255" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702841 4921 flags.go:64] FLAG: --register-node="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702846 4921 flags.go:64] FLAG: --register-schedulable="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702852 4921 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702863 4921 flags.go:64] FLAG: --registry-burst="10" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702870 4921 flags.go:64] FLAG: --registry-qps="5" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702876 4921 flags.go:64] FLAG: --reserved-cpus="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702882 4921 flags.go:64] FLAG: --reserved-memory="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702890 4921 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702897 4921 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702903 4921 flags.go:64] FLAG: --rotate-certificates="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702909 4921 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702916 4921 flags.go:64] FLAG: --runonce="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702923 4921 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702930 4921 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702937 4921 flags.go:64] FLAG: --seccomp-default="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702944 4921 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702952 4921 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702959 4921 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702966 4921 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702973 4921 flags.go:64] FLAG: --storage-driver-password="root" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702981 4921 flags.go:64] FLAG: --storage-driver-secure="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702987 4921 flags.go:64] FLAG: --storage-driver-table="stats" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702993 4921 flags.go:64] FLAG: --storage-driver-user="root" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.702999 4921 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703005 4921 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703012 4921 flags.go:64] FLAG: --system-cgroups="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703018 4921 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703031 4921 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703038 4921 flags.go:64] FLAG: --tls-cert-file="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703044 4921 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703053 4921 flags.go:64] FLAG: --tls-min-version="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703060 4921 flags.go:64] FLAG: --tls-private-key-file="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703066 4921 flags.go:64] FLAG: --topology-manager-policy="none" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703072 4921 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703078 4921 flags.go:64] FLAG: --topology-manager-scope="container" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703085 4921 flags.go:64] FLAG: --v="2" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703094 4921 flags.go:64] FLAG: --version="false" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703103 4921 flags.go:64] FLAG: --vmodule="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703111 4921 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703118 4921 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703316 4921 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703324 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703332 4921 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703339 4921 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703345 4921 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703351 4921 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703357 4921 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703363 4921 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703369 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703374 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703380 4921 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703386 4921 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703392 4921 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703398 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703403 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703409 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703415 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703421 4921 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703428 4921 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703434 4921 feature_gate.go:330] unrecognized feature gate: Example Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703439 4921 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703446 4921 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703453 4921 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703461 4921 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703467 4921 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703473 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703480 4921 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703487 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703494 4921 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703500 4921 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703506 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703513 4921 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703519 4921 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703526 4921 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703533 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703539 4921 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703553 4921 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703559 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703564 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703570 4921 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703578 4921 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703584 4921 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703589 4921 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703595 4921 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703600 4921 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703605 4921 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703611 4921 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703616 4921 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703622 4921 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703627 4921 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703632 4921 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703637 4921 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703643 4921 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703648 4921 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703654 4921 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703659 4921 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703664 4921 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703670 4921 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703676 4921 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703692 4921 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703697 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703702 4921 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703708 4921 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703713 4921 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703718 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703723 4921 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703728 4921 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703734 4921 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703740 4921 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703745 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.703750 4921 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.703767 4921 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.715433 4921 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.715492 4921 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715618 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715634 4921 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715639 4921 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715643 4921 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715648 4921 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715652 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715657 4921 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715662 4921 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715666 4921 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715671 4921 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715675 4921 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715678 4921 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715682 4921 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715686 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715700 4921 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715704 4921 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715709 4921 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715719 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715724 4921 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715727 4921 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715731 4921 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715735 4921 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715738 4921 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715742 4921 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715746 4921 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715749 4921 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715753 4921 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715756 4921 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715760 4921 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715764 4921 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715768 4921 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715772 4921 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715776 4921 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715779 4921 feature_gate.go:330] unrecognized feature gate: Example Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715783 4921 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715787 4921 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715791 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715795 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715801 4921 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715808 4921 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715813 4921 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715819 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715824 4921 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715828 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715833 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715837 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715841 4921 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715845 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715849 4921 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715854 4921 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715863 4921 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715868 4921 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715872 4921 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715876 4921 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715881 4921 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715885 4921 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715890 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715894 4921 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715897 4921 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715903 4921 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715907 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715912 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715918 4921 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715922 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715926 4921 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715930 4921 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715934 4921 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715937 4921 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715941 4921 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715944 4921 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.715948 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.715956 4921 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716108 4921 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716117 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716121 4921 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716125 4921 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716130 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716134 4921 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716138 4921 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716141 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716145 4921 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716149 4921 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716154 4921 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716158 4921 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716162 4921 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716166 4921 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716171 4921 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716176 4921 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716180 4921 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716184 4921 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716188 4921 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716192 4921 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716196 4921 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716200 4921 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716204 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716208 4921 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716212 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716216 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716221 4921 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716225 4921 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716230 4921 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716234 4921 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716238 4921 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716242 4921 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716246 4921 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716250 4921 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716255 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716259 4921 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716263 4921 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716282 4921 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716286 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716291 4921 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716296 4921 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716301 4921 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716306 4921 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716311 4921 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716315 4921 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716321 4921 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716326 4921 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716330 4921 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716334 4921 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716338 4921 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716343 4921 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716346 4921 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716350 4921 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716354 4921 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716357 4921 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716361 4921 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716365 4921 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716368 4921 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716373 4921 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716378 4921 feature_gate.go:330] unrecognized feature gate: Example Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716382 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716386 4921 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716390 4921 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716393 4921 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716397 4921 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716400 4921 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716405 4921 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716410 4921 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716414 4921 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716418 4921 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.716423 4921 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.716429 4921 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.716672 4921 server.go:940] "Client rotation is on, will bootstrap in background" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.719900 4921 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.720005 4921 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.720708 4921 server.go:997] "Starting client certificate rotation" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.720732 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.720943 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-13 20:13:12.219071547 +0000 UTC Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.721075 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.728940 4921 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.733071 4921 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.733734 4921 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.749187 4921 log.go:25] "Validated CRI v1 runtime API" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.767197 4921 log.go:25] "Validated CRI v1 image API" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.768887 4921 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.771520 4921 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-03-03-37-33-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.771550 4921 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.784107 4921 manager.go:217] Machine: {Timestamp:2026-01-03 03:41:04.782953744 +0000 UTC m=+0.394380578 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3994d32f-6f26-48b7-a835-33b55e41bd4d BootID:904f1a66-3c4f-4ba4-86d9-b5f3204a6dea Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a7:b3:65 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a7:b3:65 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5f:fe:14 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3b:be:e4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:61:4a:96 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:60:7a:1c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d6:67:c2:31:29:e7 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:3d:92:03:0a:11 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.784340 4921 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.784452 4921 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785020 4921 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785170 4921 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785204 4921 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785434 4921 topology_manager.go:138] "Creating topology manager with none policy" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785444 4921 container_manager_linux.go:303] "Creating device plugin manager" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785602 4921 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785629 4921 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.785926 4921 state_mem.go:36] "Initialized new in-memory state store" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.786459 4921 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.787479 4921 kubelet.go:418] "Attempting to sync node with API server" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.787503 4921 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.787616 4921 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.787632 4921 kubelet.go:324] "Adding apiserver pod source" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.787644 4921 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.789771 4921 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.790385 4921 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.791112 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.793437 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.793538 4921 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794390 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794437 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794466 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794482 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794508 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794523 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794538 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794563 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794582 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794597 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794623 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.794638 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.795229 4921 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.796313 4921 server.go:1280] "Started kubelet" Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.797035 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.797617 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.798070 4921 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.798003 4921 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 03 03:41:04 crc systemd[1]: Started Kubernetes Kubelet. Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.799672 4921 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.801036 4921 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.803909 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.803966 4921 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.804056 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 03:57:21.510146049 +0000 UTC Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.804524 4921 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.804556 4921 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.804641 4921 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.804745 4921 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.805977 4921 factory.go:55] Registering systemd factory Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806031 4921 factory.go:221] Registration of the systemd container factory successfully Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.805972 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806472 4921 factory.go:153] Registering CRI-O factory Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806497 4921 factory.go:221] Registration of the crio container factory successfully Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806621 4921 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806685 4921 factory.go:103] Registering Raw factory Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.806711 4921 manager.go:1196] Started watching for new ooms in manager Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.806718 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.806804 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.807760 4921 manager.go:319] Starting recovery of all containers Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.805794 4921 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18871b8b840fc6c4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-03 03:41:04.796239556 +0000 UTC m=+0.407666410,LastTimestamp:2026-01-03 03:41:04.796239556 +0000 UTC m=+0.407666410,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.815181 4921 server.go:460] "Adding debug handlers to kubelet server" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.824304 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826070 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826098 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826119 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826146 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826161 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826183 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826200 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826239 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826263 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826329 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826358 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826376 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826424 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826441 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826461 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826480 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826626 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826700 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826726 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826752 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826852 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.826954 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827003 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827072 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827102 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827207 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827329 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827381 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827413 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.827957 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828021 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828057 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828101 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828130 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828169 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828185 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828210 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828287 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828305 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828387 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828442 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828492 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828536 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828566 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828695 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828769 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828797 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828815 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828957 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.828985 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829023 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829159 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829201 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829312 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829343 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829389 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829419 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829746 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829821 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829841 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.829890 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830095 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830115 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830343 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830491 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830522 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830546 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830603 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830650 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830734 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830761 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830784 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830836 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830910 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830928 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830944 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.830997 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831039 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831084 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831100 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831195 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831256 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831298 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831323 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831344 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831404 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831428 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831462 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831534 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831653 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831670 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831686 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831697 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831778 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831850 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831887 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831911 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.831933 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832017 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832071 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832084 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832137 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832153 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832280 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832389 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832428 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832446 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832468 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832482 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832531 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832557 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832582 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832605 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832621 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832644 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832665 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832709 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832739 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832754 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832769 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832795 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832833 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832854 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832913 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832942 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.832987 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833047 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833130 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833145 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833177 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833240 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833299 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833369 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833403 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833416 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833453 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833463 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833517 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833530 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833541 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833556 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833640 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833655 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833677 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833689 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833704 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833713 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833751 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833764 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833808 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833822 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833908 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833949 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.833977 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834020 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834056 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834072 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834108 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834126 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834189 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834205 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834218 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834235 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834290 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834306 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834342 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834361 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834375 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834462 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834504 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834519 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834540 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.834552 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835340 4921 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835383 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835396 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835407 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835436 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835447 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835477 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835527 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835552 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835572 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835627 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835640 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835672 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835682 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835757 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835800 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835855 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835893 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835904 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835919 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835931 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835970 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.835985 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836000 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836059 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836076 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836086 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836100 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836145 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836162 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836174 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836229 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836260 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836294 4921 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836303 4921 reconstruct.go:97] "Volume reconstruction finished" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.836310 4921 reconciler.go:26] "Reconciler: start to sync state" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.840176 4921 manager.go:324] Recovery completed Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.859857 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.862761 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.862810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.862821 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.864542 4921 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.864575 4921 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.864595 4921 state_mem.go:36] "Initialized new in-memory state store" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.874683 4921 policy_none.go:49] "None policy: Start" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.876597 4921 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.876638 4921 state_mem.go:35] "Initializing new in-memory state store" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.879078 4921 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.882408 4921 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.882460 4921 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.882490 4921 kubelet.go:2335] "Starting kubelet main sync loop" Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.882547 4921 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 03 03:41:04 crc kubenswrapper[4921]: W0103 03:41:04.884072 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.884188 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.904854 4921 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.938198 4921 manager.go:334] "Starting Device Plugin manager" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.938372 4921 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.938398 4921 server.go:79] "Starting device plugin registration server" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.939241 4921 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.939314 4921 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.939693 4921 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.939864 4921 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.939878 4921 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 03 03:41:04 crc kubenswrapper[4921]: E0103 03:41:04.948620 4921 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.983514 4921 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.983671 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985242 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985257 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985413 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985620 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.985689 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986649 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986665 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986719 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986739 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986759 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986886 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.986991 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.987046 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988299 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988333 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988347 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988678 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988893 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.988991 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.989434 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.989479 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.989499 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990325 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990358 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990421 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990440 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990718 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990860 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.990925 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992298 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992329 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992332 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992364 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992385 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992342 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992645 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.992674 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.994087 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.994135 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:04 crc kubenswrapper[4921]: I0103 03:41:04.994162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.007950 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.038792 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.038838 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.038871 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.038894 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.038980 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039166 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039295 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039349 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039398 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039446 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039481 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039527 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039570 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039640 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039710 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.039816 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.040564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.040612 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.040624 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.040654 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.041206 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141021 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141106 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141155 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141190 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141301 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141317 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141381 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141331 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141435 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141444 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141467 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141531 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141538 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141576 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141584 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141614 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141623 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141648 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141668 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141682 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141707 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141712 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141745 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141769 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141797 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141837 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141878 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141942 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.141989 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.142035 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.241369 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.243299 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.243358 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.243378 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.243451 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.243951 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.317338 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.324255 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.347156 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.356357 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.361122 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:05 crc kubenswrapper[4921]: W0103 03:41:05.377597 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3e2be187c035fab070aac6df72febf1df1bdedda765ced09733ed44fb5f8a5de WatchSource:0}: Error finding container 3e2be187c035fab070aac6df72febf1df1bdedda765ced09733ed44fb5f8a5de: Status 404 returned error can't find the container with id 3e2be187c035fab070aac6df72febf1df1bdedda765ced09733ed44fb5f8a5de Jan 03 03:41:05 crc kubenswrapper[4921]: W0103 03:41:05.382012 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-99e34c4f93276f727439af7cee153ba221fc2da87ce5dc45a39bc50f1d377787 WatchSource:0}: Error finding container 99e34c4f93276f727439af7cee153ba221fc2da87ce5dc45a39bc50f1d377787: Status 404 returned error can't find the container with id 99e34c4f93276f727439af7cee153ba221fc2da87ce5dc45a39bc50f1d377787 Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.409112 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.644431 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.646176 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.646251 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.646284 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.646326 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.647064 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.802439 4921 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.805559 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 18:34:54.003155425 +0000 UTC Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.894340 4921 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e" exitCode=0 Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.894439 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.894595 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b38bbf46b7cc0b9f45d942965c7b7e5e20ef26ca49c9aa2091066575aefa3c06"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.894757 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.899051 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.899091 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.899103 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.900622 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.900672 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8c6f23ac3439e732f9d602a5e8120a46f404fcc815a4db917411623850a4fb0d"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.902709 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21" exitCode=0 Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.902780 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.902811 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"99e34c4f93276f727439af7cee153ba221fc2da87ce5dc45a39bc50f1d377787"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.902938 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.904633 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.904666 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.904677 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.905463 4921 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="34dbfd745a8ad18537b340975b2eb8aed5405d5721c85363900b1446570e1744" exitCode=0 Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.905524 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"34dbfd745a8ad18537b340975b2eb8aed5405d5721c85363900b1446570e1744"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.905541 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3e2be187c035fab070aac6df72febf1df1bdedda765ced09733ed44fb5f8a5de"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.905618 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.906387 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.906554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.906590 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.906602 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907023 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907045 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907582 4921 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="07420fbd9171cf1d81df4435de83ae2fef398960174883db6caf28a15d97e0c0" exitCode=0 Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907628 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"07420fbd9171cf1d81df4435de83ae2fef398960174883db6caf28a15d97e0c0"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907646 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"48639297275e2ae0aacf0b849184cb477fd34b0d659c5f162e11ef726a0ee58d"} Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.907722 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.908332 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.908351 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:05 crc kubenswrapper[4921]: I0103 03:41:05.908359 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:05 crc kubenswrapper[4921]: W0103 03:41:05.977894 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:05 crc kubenswrapper[4921]: E0103 03:41:05.978068 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:06 crc kubenswrapper[4921]: W0103 03:41:06.121341 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.121417 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.210204 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s" Jan 03 03:41:06 crc kubenswrapper[4921]: W0103 03:41:06.274217 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.274364 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:06 crc kubenswrapper[4921]: W0103 03:41:06.297738 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.297839 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.447702 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.449404 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.449452 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.449466 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.449496 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.450206 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.251:6443: connect: connection refused" node="crc" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.802858 4921 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.251:6443: connect: connection refused Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.805889 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 19:41:23.762595062 +0000 UTC Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.905771 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 03 03:41:06 crc kubenswrapper[4921]: E0103 03:41:06.907060 4921 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.251:6443: connect: connection refused" logger="UnhandledError" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.911783 4921 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="015d2cc40449bafa241d2730a5482c218a2a582b354f30e569ea9b029f571185" exitCode=0 Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.911853 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"015d2cc40449bafa241d2730a5482c218a2a582b354f30e569ea9b029f571185"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.911994 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.913647 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.913702 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.913715 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.914873 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"30f197eb78a36e25db364d4946e777a61b691206bc26226d7aa9142ab2e3310c"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.914945 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.916830 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.916854 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.916864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.919239 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.919305 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.919324 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.919406 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.920547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.920576 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.920588 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.928412 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.928471 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.928487 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.928591 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.929661 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.929701 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.929716 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.938587 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.938676 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.938696 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954"} Jan 03 03:41:06 crc kubenswrapper[4921]: I0103 03:41:06.938714 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30"} Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.806292 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 16:34:21.355381827 +0000 UTC Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.946376 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd"} Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.946481 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.947745 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.947807 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.947824 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.949615 4921 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="39c982cc65789752e348e9c82e1560f859db01cfdcae446381b73a3fe937e31e" exitCode=0 Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.949683 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"39c982cc65789752e348e9c82e1560f859db01cfdcae446381b73a3fe937e31e"} Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.949792 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.949852 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.954228 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955005 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955066 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955149 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955172 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.955182 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.956743 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.956840 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:07 crc kubenswrapper[4921]: I0103 03:41:07.956905 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.050561 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.052096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.052143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.052155 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.052188 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.737693 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.738036 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.740039 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.740110 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.740130 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.806673 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 19:56:42.588957827 +0000 UTC Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.957846 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"744023d6ed4cfda772e77fd5a85c21cd7b803475ee493f55edd4f25da26b5188"} Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.957923 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c513ef96f979d66631578ae812e6dbc94ec2a7baecdf2d51d4a1257d97fd30a"} Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.957945 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.957947 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"00220682f9af2888703498c023213a3edca0eeec42eab348653ef37a93f3cfab"} Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.958084 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.961195 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.961251 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:08 crc kubenswrapper[4921]: I0103 03:41:08.961281 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.264681 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.264890 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.267353 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.267408 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.267428 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.574626 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.807761 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 03:28:31.210022896 +0000 UTC Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.965949 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d9482a194b5147fb85f2c284603e5fa83832b12b4cea2a2aa0d780f7884ccf2c"} Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.966011 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7bf14f56b361b7486eed6790758199e36d482e833965047cdaa04abb287d3b73"} Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.966029 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.966216 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.966963 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.967004 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.967017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.967858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.967892 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:09 crc kubenswrapper[4921]: I0103 03:41:09.967902 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.304313 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.361702 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.361920 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.367056 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.367121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.367144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.700976 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.808769 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 22:18:37.935187002 +0000 UTC Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.969618 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.969728 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.970970 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.971048 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.971075 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.971199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.971244 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:10 crc kubenswrapper[4921]: I0103 03:41:10.971259 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.103722 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.809075 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 05:15:44.144211057 +0000 UTC Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.973258 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.974778 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.974827 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:11 crc kubenswrapper[4921]: I0103 03:41:11.974844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.661612 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.662220 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.664387 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.664581 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.664721 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.670131 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.696500 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.696963 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.698469 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.698530 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.698548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.809870 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 12:45:14.437719147 +0000 UTC Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.976313 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.978060 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.978328 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:12 crc kubenswrapper[4921]: I0103 03:41:12.978556 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:13 crc kubenswrapper[4921]: I0103 03:41:13.811351 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 06:08:35.642477196 +0000 UTC Jan 03 03:41:13 crc kubenswrapper[4921]: I0103 03:41:13.811457 4921 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 146h27m21.831025063s for next certificate rotation Jan 03 03:41:14 crc kubenswrapper[4921]: E0103 03:41:14.948963 4921 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.493055 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.493310 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.496527 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.496593 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.496916 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.498696 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.985707 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.987089 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.987177 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:15 crc kubenswrapper[4921]: I0103 03:41:15.987200 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:17 crc kubenswrapper[4921]: I0103 03:41:17.494165 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 03 03:41:17 crc kubenswrapper[4921]: I0103 03:41:17.494561 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 03 03:41:17 crc kubenswrapper[4921]: W0103 03:41:17.702678 4921 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Jan 03 03:41:17 crc kubenswrapper[4921]: I0103 03:41:17.702810 4921 trace.go:236] Trace[985904845]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Jan-2026 03:41:07.701) (total time: 10001ms): Jan 03 03:41:17 crc kubenswrapper[4921]: Trace[985904845]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:41:17.702) Jan 03 03:41:17 crc kubenswrapper[4921]: Trace[985904845]: [10.001294131s] [10.001294131s] END Jan 03 03:41:17 crc kubenswrapper[4921]: E0103 03:41:17.702846 4921 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 03 03:41:17 crc kubenswrapper[4921]: I0103 03:41:17.803699 4921 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 03 03:41:17 crc kubenswrapper[4921]: E0103 03:41:17.811195 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Jan 03 03:41:18 crc kubenswrapper[4921]: E0103 03:41:18.053814 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Jan 03 03:41:18 crc kubenswrapper[4921]: E0103 03:41:18.169308 4921 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.18871b8b840fc6c4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-03 03:41:04.796239556 +0000 UTC m=+0.407666410,LastTimestamp:2026-01-03 03:41:04.796239556 +0000 UTC m=+0.407666410,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.277918 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.277999 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.284855 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.284915 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.493948 4921 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 03 03:41:18 crc kubenswrapper[4921]: I0103 03:41:18.494061 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.333878 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.334059 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.335616 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.335713 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.335734 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.349158 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 03 03:41:20 crc kubenswrapper[4921]: I0103 03:41:20.998854 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.000516 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.000579 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.000600 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.254782 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.256188 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.256291 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.256312 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:21 crc kubenswrapper[4921]: I0103 03:41:21.256359 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:21 crc kubenswrapper[4921]: E0103 03:41:21.262073 4921 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.703548 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.703837 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.705522 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.705593 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.705612 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:22 crc kubenswrapper[4921]: I0103 03:41:22.710669 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.004654 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.006158 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.006214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.006226 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.266381 4921 trace.go:236] Trace[1938091395]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Jan-2026 03:41:08.839) (total time: 14426ms): Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[1938091395]: ---"Objects listed" error: 14426ms (03:41:23.266) Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[1938091395]: [14.426424071s] [14.426424071s] END Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.266420 4921 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.268487 4921 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.269252 4921 trace.go:236] Trace[433440799]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Jan-2026 03:41:08.738) (total time: 14531ms): Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[433440799]: ---"Objects listed" error: 14531ms (03:41:23.269) Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[433440799]: [14.531083608s] [14.531083608s] END Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.269291 4921 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.269735 4921 trace.go:236] Trace[597009128]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Jan-2026 03:41:09.281) (total time: 13987ms): Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[597009128]: ---"Objects listed" error: 13987ms (03:41:23.269) Jan 03 03:41:23 crc kubenswrapper[4921]: Trace[597009128]: [13.987920197s] [13.987920197s] END Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.269780 4921 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.272347 4921 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.307066 4921 csr.go:261] certificate signing request csr-nsmxg is approved, waiting to be issued Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.316366 4921 csr.go:257] certificate signing request csr-nsmxg is issued Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.406511 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57966->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.406608 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57966->192.168.126.11:17697: read: connection reset by peer" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.407113 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.407207 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.470061 4921 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.800375 4921 apiserver.go:52] "Watching apiserver" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.805179 4921 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.807051 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-qnm8g","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.807576 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.807746 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.807771 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.807596 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:23 crc kubenswrapper[4921]: E0103 03:41:23.807872 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:23 crc kubenswrapper[4921]: E0103 03:41:23.807940 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.808042 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.808258 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.808417 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:23 crc kubenswrapper[4921]: E0103 03:41:23.808627 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.810984 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.811089 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.811192 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.811086 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.811285 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.813878 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.814052 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.814212 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.815404 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.815764 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-mhxl4"] Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.816078 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.817025 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.820229 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.823915 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.823926 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.824050 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.824046 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.825887 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.852053 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.871220 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.887418 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.900737 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.906538 4921 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.912493 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.924164 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.934391 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.944528 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.952561 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.961412 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972849 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972896 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972919 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972939 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972959 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.972978 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973003 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973019 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973038 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973055 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973072 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973089 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973104 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973119 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973134 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973152 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973175 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973195 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973214 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973229 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973247 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973287 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.973312 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974144 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974219 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974259 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974245 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974315 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974237 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974360 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974399 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974393 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974430 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974459 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974584 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974606 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974760 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974844 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974896 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974944 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975072 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975094 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975176 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975210 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975359 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975421 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975483 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975495 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975634 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975797 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975800 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975830 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975878 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975908 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975905 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.975906 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.974482 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976087 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976136 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976148 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976159 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976171 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976261 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976327 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976372 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976449 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976487 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976517 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976523 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976520 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976582 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976607 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976633 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976661 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976747 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976758 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976781 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976808 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976777 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976877 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976918 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976952 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976990 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977034 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977067 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977100 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977141 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977179 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977213 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977249 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977300 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977330 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977364 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977393 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977419 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977449 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977489 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977523 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977552 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977585 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977624 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977663 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977695 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977730 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977766 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977805 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977840 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977876 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977915 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977955 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977993 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978036 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978074 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978109 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978256 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978954 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979006 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979043 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979084 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979125 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979166 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979206 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979242 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979302 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979341 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979490 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979559 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979613 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979647 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979695 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979744 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979775 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979809 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979852 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979904 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979941 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979985 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.980029 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.980064 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984519 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984738 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.976990 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977088 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977448 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977474 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977565 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977892 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.977932 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978124 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978169 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978262 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978539 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.978835 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979062 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979079 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979304 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979374 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979518 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979547 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979613 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979811 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979828 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979885 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.979969 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.980098 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.980129 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.980522 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.981009 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.981067 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.981388 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.981436 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.981891 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982433 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982436 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982450 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982650 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982670 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982846 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982965 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.983059 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.982993 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.983100 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.983135 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984062 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984209 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984223 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984240 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984619 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984515 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.984845 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985316 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985496 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985614 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985624 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: E0103 03:41:23.985677 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:24.484993018 +0000 UTC m=+20.096419842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985669 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985895 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.985969 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.987198 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.987307 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.987336 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.987363 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.988430 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.988781 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.988971 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989003 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989056 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989085 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989110 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989138 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989160 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989198 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989223 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989251 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989294 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989321 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989342 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989415 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989444 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989466 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989498 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989525 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989545 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989563 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989587 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989607 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989625 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989647 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989666 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989683 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989702 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989721 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989744 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989760 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989780 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989808 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989826 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989847 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989867 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989887 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989906 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989927 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989948 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989968 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989988 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.989982 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990008 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990030 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990040 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990050 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990142 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990179 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990217 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990253 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990783 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990889 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990933 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990936 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.990990 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.991017 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.992844 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.992904 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.992945 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.992953 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.992988 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993019 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993012 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993060 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993094 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993127 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993154 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993163 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993171 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993184 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993314 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993374 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993405 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993437 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993460 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993483 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993503 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993526 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993547 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993573 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993597 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993611 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993621 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993687 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993726 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993733 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993754 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993770 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993807 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993838 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993868 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993893 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993932 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993962 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.993996 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994021 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994096 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994125 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994202 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994408 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994484 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994609 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994795 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995121 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995202 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.994127 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995847 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995943 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995980 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996007 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996036 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996077 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996110 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-hosts-file\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996142 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996171 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c62912da-e574-4877-9c1a-59e2a10c1d18-host\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996192 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qtxx\" (UniqueName: \"kubernetes.io/projected/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-kube-api-access-5qtxx\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996218 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996241 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996325 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996350 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996375 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996433 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996459 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996482 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996511 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c62912da-e574-4877-9c1a-59e2a10c1d18-serviceca\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996533 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ck4\" (UniqueName: \"kubernetes.io/projected/c62912da-e574-4877-9c1a-59e2a10c1d18-kube-api-access-t4ck4\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996650 4921 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996669 4921 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996686 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996700 4921 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996714 4921 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996727 4921 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996743 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996757 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996772 4921 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996786 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996801 4921 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996813 4921 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996825 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996837 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996860 4921 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996877 4921 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996897 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996910 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996923 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996936 4921 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996949 4921 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996967 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996982 4921 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996996 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997011 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997024 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997035 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997047 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997059 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997071 4921 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997082 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997093 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997104 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997115 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997127 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997143 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997154 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997167 4921 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997181 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997193 4921 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997205 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997220 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997234 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997250 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997262 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997290 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997302 4921 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997313 4921 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997326 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997340 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997354 4921 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997368 4921 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997379 4921 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997392 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997404 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997416 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997428 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997439 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997449 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997461 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997474 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997488 4921 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997502 4921 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997514 4921 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997527 4921 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997539 4921 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997551 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997564 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997577 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997625 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997640 4921 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997653 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997666 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997679 4921 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997692 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997705 4921 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997719 4921 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997731 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997743 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997756 4921 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997770 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997782 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997796 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997809 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997822 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997834 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997847 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997862 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997875 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997893 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997906 4921 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997920 4921 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997931 4921 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997943 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997955 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997968 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997979 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997990 4921 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998003 4921 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998015 4921 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998027 4921 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998038 4921 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998050 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998061 4921 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998073 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998086 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998116 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998134 4921 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998147 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998159 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998183 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.995427 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996335 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996676 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996699 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.996731 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997060 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997095 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.997591 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998065 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.998253 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.999478 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:23 crc kubenswrapper[4921]: I0103 03:41:23.999686 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000353 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000443 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000506 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000757 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000853 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000933 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.000980 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.001127 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.002966 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:24.502939635 +0000 UTC m=+20.114366479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002973 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.001165 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002555 4921 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.001398 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003002 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.001541 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.001813 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002105 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002534 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.002701 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002760 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.002801 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003142 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003173 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003427 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003502 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003715 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.003812 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.004083 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.004105 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.004496 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.004521 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.004877 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.004896 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:24.504863077 +0000 UTC m=+20.116290071 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.019928 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.022741 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.022773 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.022791 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.022868 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:24.522846936 +0000 UTC m=+20.134273780 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.027930 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.028985 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.035600 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.039010 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.039335 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.039374 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.039617 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.039636 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.039650 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.039707 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:24.539686414 +0000 UTC m=+20.151113228 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.040213 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.040296 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.040474 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.040538 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.041357 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.044455 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.045803 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.046640 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.047621 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.047877 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.047971 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.048005 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.048783 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.049966 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.050175 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.050354 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.055115 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.055410 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.056617 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.056710 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.057231 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.057246 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.057293 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.057479 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.058286 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.058611 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.059875 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.060962 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.062050 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.062127 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd" exitCode=255 Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.062205 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd"} Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.069714 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.079867 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.086484 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.086836 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.087804 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.092081 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098085 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098690 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c62912da-e574-4877-9c1a-59e2a10c1d18-host\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098734 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qtxx\" (UniqueName: \"kubernetes.io/projected/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-kube-api-access-5qtxx\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098763 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098811 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c62912da-e574-4877-9c1a-59e2a10c1d18-serviceca\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098836 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ck4\" (UniqueName: \"kubernetes.io/projected/c62912da-e574-4877-9c1a-59e2a10c1d18-kube-api-access-t4ck4\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098872 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098895 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-hosts-file\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098939 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098954 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098969 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098981 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.098994 4921 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099006 4921 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099019 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099032 4921 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099046 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099060 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099072 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099084 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099097 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099112 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099125 4921 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099137 4921 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099148 4921 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099160 4921 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099172 4921 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099186 4921 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099199 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099210 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099224 4921 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099236 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099247 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099260 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099292 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099305 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099318 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099367 4921 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099382 4921 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099396 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099409 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099421 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099432 4921 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099448 4921 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099461 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099474 4921 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099486 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099498 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099510 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099521 4921 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099536 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099549 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099565 4921 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099578 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099589 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099600 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099611 4921 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099622 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099634 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099645 4921 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099658 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099669 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099680 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099692 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099702 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099713 4921 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099726 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099737 4921 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099748 4921 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099759 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099774 4921 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099786 4921 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099797 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099807 4921 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099819 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099832 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099844 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099855 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099867 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099878 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099891 4921 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.100178 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.100233 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c62912da-e574-4877-9c1a-59e2a10c1d18-host\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.102501 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.102620 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-hosts-file\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.102833 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c62912da-e574-4877-9c1a-59e2a10c1d18-serviceca\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.099902 4921 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.102947 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.102963 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.103164 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.103190 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.103459 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.103715 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.108134 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.109829 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.114638 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.121117 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.127749 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.128963 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.129841 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.133827 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qtxx\" (UniqueName: \"kubernetes.io/projected/eed39a0f-cb1e-451a-a36c-233b20b3ca1a-kube-api-access-5qtxx\") pod \"node-resolver-qnm8g\" (UID: \"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\") " pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.137211 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ck4\" (UniqueName: \"kubernetes.io/projected/c62912da-e574-4877-9c1a-59e2a10c1d18-kube-api-access-t4ck4\") pod \"node-ca-mhxl4\" (UID: \"c62912da-e574-4877-9c1a-59e2a10c1d18\") " pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.138410 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qnm8g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.145867 4921 scope.go:117] "RemoveContainer" containerID="2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.146482 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.148463 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.151286 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.154586 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-mhxl4" Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.179210 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-32fa1ab5fb007de8f7a4bde1966442637e43e2c92aba19d69d8547df61db35d6 WatchSource:0}: Error finding container 32fa1ab5fb007de8f7a4bde1966442637e43e2c92aba19d69d8547df61db35d6: Status 404 returned error can't find the container with id 32fa1ab5fb007de8f7a4bde1966442637e43e2c92aba19d69d8547df61db35d6 Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.199244 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.202169 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc62912da_e574_4877_9c1a_59e2a10c1d18.slice/crio-0295974055bea48df613818770c47872dc587a7d1f1c44c872d9ae4f645004f7 WatchSource:0}: Error finding container 0295974055bea48df613818770c47872dc587a7d1f1c44c872d9ae4f645004f7: Status 404 returned error can't find the container with id 0295974055bea48df613818770c47872dc587a7d1f1c44c872d9ae4f645004f7 Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205729 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205758 4921 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205770 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205783 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205794 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205803 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.205817 4921 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.217105 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.239316 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.320253 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-03 03:36:23 +0000 UTC, rotation deadline is 2026-10-22 23:39:35.455271308 +0000 UTC Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.320357 4921 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7027h58m11.134917646s for next certificate rotation Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.509052 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.509173 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.509210 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.509288 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:25.509242835 +0000 UTC m=+21.120669659 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.509360 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.509387 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.509437 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:25.509416259 +0000 UTC m=+21.120843083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.509460 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:25.50945055 +0000 UTC m=+21.120877374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.587130 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-6xvww"] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.588055 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-cctxw"] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.588293 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.588454 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.592723 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.592802 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.592968 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593158 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593229 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593236 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbb7k"] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593283 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593486 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593333 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.593621 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.594253 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-666ct"] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.594409 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.594605 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.597528 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.598509 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.598707 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.599559 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.599770 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.600235 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.600637 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.600719 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.600955 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.603577 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.608768 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.609664 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.609835 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.609869 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.609883 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.609945 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:25.609924707 +0000 UTC m=+21.221351531 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.610065 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.610203 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.610243 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.610261 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: E0103 03:41:24.610359 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:25.610333098 +0000 UTC m=+21.221760112 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.626305 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.645264 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.659298 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.674151 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.687382 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.704763 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711100 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711162 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711185 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711207 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrslm\" (UniqueName: \"kubernetes.io/projected/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-kube-api-access-mrslm\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711247 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5sd\" (UniqueName: \"kubernetes.io/projected/429ab47e-68f8-4b60-aa4c-ab79a764b7db-kube-api-access-xm5sd\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711262 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711301 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711319 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cnibin\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711335 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711372 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711391 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711417 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711447 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711465 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711482 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-hostroot\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711497 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-daemon-config\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711528 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-k8s-cni-cncf-io\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711543 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711562 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711578 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-os-release\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711610 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-multus\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711625 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-os-release\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711641 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-binary-copy\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711681 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711698 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711711 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711733 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711762 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711776 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-kubelet\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711797 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-bin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711811 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/429ab47e-68f8-4b60-aa4c-ab79a764b7db-mcd-auth-proxy-config\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711850 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711886 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-netns\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711924 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/429ab47e-68f8-4b60-aa4c-ab79a764b7db-proxy-tls\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.711946 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712002 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgv5c\" (UniqueName: \"kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712018 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cni-binary-copy\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712044 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/429ab47e-68f8-4b60-aa4c-ab79a764b7db-rootfs\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712084 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-system-cni-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712106 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cnibin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712124 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-socket-dir-parent\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712159 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-conf-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712175 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-multus-certs\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712189 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6xb\" (UniqueName: \"kubernetes.io/projected/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-kube-api-access-9t6xb\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712211 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712239 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-system-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712253 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-etc-kubernetes\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.712299 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.718133 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.721898 4921 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722183 4921 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722248 4921 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722318 4921 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722366 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722416 4921 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722457 4921 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722503 4921 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722542 4921 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722586 4921 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722626 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722666 4921 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.722709 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.723845 4921 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724126 4921 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724166 4921 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724202 4921 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724235 4921 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724300 4921 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724345 4921 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724385 4921 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724431 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724478 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724540 4921 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724600 4921 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724538 4921 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724667 4921 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724718 4921 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724767 4921 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724801 4921 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724881 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724930 4921 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.724975 4921 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.725014 4921 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.725050 4921 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.725137 4921 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.813733 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814331 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814423 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814426 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814551 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-daemon-config\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814580 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-hostroot\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814599 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-k8s-cni-cncf-io\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814623 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-multus\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814647 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-os-release\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814686 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-binary-copy\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814709 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814737 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814731 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-multus\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814754 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-os-release\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814784 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814783 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-hostroot\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814803 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814936 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-os-release\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814936 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814979 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-os-release\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814986 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815020 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815061 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-kubelet\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815135 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-bin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815166 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/429ab47e-68f8-4b60-aa4c-ab79a764b7db-mcd-auth-proxy-config\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815206 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815237 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-netns\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815300 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/429ab47e-68f8-4b60-aa4c-ab79a764b7db-proxy-tls\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815336 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cni-binary-copy\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815368 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/429ab47e-68f8-4b60-aa4c-ab79a764b7db-rootfs\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815401 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-system-cni-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815436 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815473 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgv5c\" (UniqueName: \"kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815504 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-socket-dir-parent\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.814558 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815538 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-conf-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815572 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-multus-certs\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815585 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815607 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6xb\" (UniqueName: \"kubernetes.io/projected/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-kube-api-access-9t6xb\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815648 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cnibin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815684 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-system-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815719 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-binary-copy\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815720 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-etc-kubernetes\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815766 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-etc-kubernetes\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815832 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815874 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815906 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815897 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-daemon-config\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815967 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.815933 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816009 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816047 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrslm\" (UniqueName: \"kubernetes.io/projected/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-kube-api-access-mrslm\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816074 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816098 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816113 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-k8s-cni-cncf-io\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816133 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5sd\" (UniqueName: \"kubernetes.io/projected/429ab47e-68f8-4b60-aa4c-ab79a764b7db-kube-api-access-xm5sd\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816146 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816052 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816170 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816208 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816242 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816297 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cnibin\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816343 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816378 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816456 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.813929 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816510 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816527 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816806 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-system-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816838 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-cni-bin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.816948 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817295 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-cnibin\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817346 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817375 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-conf-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817418 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cni-binary-copy\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817469 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-cnibin\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817520 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-multus-certs\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817548 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-run-netns\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817572 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817588 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817628 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-cni-dir\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817650 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-host-var-lib-kubelet\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817967 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/429ab47e-68f8-4b60-aa4c-ab79a764b7db-mcd-auth-proxy-config\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817973 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-system-cni-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.817992 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.818020 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.818090 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/429ab47e-68f8-4b60-aa4c-ab79a764b7db-rootfs\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.818159 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.818249 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-multus-socket-dir-parent\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.818373 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.819192 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.819890 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/429ab47e-68f8-4b60-aa4c-ab79a764b7db-proxy-tls\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.837977 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5sd\" (UniqueName: \"kubernetes.io/projected/429ab47e-68f8-4b60-aa4c-ab79a764b7db-kube-api-access-xm5sd\") pod \"machine-config-daemon-cctxw\" (UID: \"429ab47e-68f8-4b60-aa4c-ab79a764b7db\") " pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.838987 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgv5c\" (UniqueName: \"kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c\") pod \"ovnkube-node-zbb7k\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.839361 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6xb\" (UniqueName: \"kubernetes.io/projected/165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f-kube-api-access-9t6xb\") pod \"multus-additional-cni-plugins-6xvww\" (UID: \"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\") " pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.841911 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrslm\" (UniqueName: \"kubernetes.io/projected/116d8b0a-baa0-4087-9a34-8e890ab8f8dc-kube-api-access-mrslm\") pod \"multus-666ct\" (UID: \"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\") " pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.887317 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.887867 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.904889 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-6xvww" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.911990 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.922114 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.928834 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-666ct" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.948454 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.955481 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.956175 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: I0103 03:41:24.957482 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.969236 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod165b4d4c_dd8c_4c0f_8a6c_8282a934fb4f.slice/crio-7f74c24b21e0e98dfa2cad0d63178fab428b9c252845a89ae56a5419381c6a5c WatchSource:0}: Error finding container 7f74c24b21e0e98dfa2cad0d63178fab428b9c252845a89ae56a5419381c6a5c: Status 404 returned error can't find the container with id 7f74c24b21e0e98dfa2cad0d63178fab428b9c252845a89ae56a5419381c6a5c Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.972236 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf34943c_bfe4_4411_af8a_189b14e35a82.slice/crio-59afa832eb9a37573ea31b7f77d861097a8dd7a8d8b4d635afa837e88844cc37 WatchSource:0}: Error finding container 59afa832eb9a37573ea31b7f77d861097a8dd7a8d8b4d635afa837e88844cc37: Status 404 returned error can't find the container with id 59afa832eb9a37573ea31b7f77d861097a8dd7a8d8b4d635afa837e88844cc37 Jan 03 03:41:24 crc kubenswrapper[4921]: W0103 03:41:24.974728 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod429ab47e_68f8_4b60_aa4c_ab79a764b7db.slice/crio-737f652182fd4bd2803099832c1c5c2b26df3a64a4e4ad4e16a4451b8d3b080a WatchSource:0}: Error finding container 737f652182fd4bd2803099832c1c5c2b26df3a64a4e4ad4e16a4451b8d3b080a: Status 404 returned error can't find the container with id 737f652182fd4bd2803099832c1c5c2b26df3a64a4e4ad4e16a4451b8d3b080a Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.015140 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.022349 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.023080 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.023858 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.027283 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.075779 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.092973 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.093593 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.096402 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.098348 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.099743 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.100538 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.101074 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.102287 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.102911 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.103435 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.104648 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.105288 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.111889 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.114730 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.115399 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.117090 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.118376 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.119735 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.120321 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.120998 4921 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.121105 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.123693 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.124494 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.126134 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.128641 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.129583 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.131361 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.132310 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.133862 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.134553 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.135462 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.138410 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.140021 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.140772 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.143392 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.144232 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.146058 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.146853 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.148019 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.149171 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.150039 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.151688 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.152376 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153600 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qnm8g" event={"ID":"eed39a0f-cb1e-451a-a36c-233b20b3ca1a","Type":"ContainerStarted","Data":"c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153671 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qnm8g" event={"ID":"eed39a0f-cb1e-451a-a36c-233b20b3ca1a","Type":"ContainerStarted","Data":"17889d710e0700c3dfc7777d226e26f6a16df834d5a7ef3b91dd5a0426ebb338"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153709 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153728 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"59afa832eb9a37573ea31b7f77d861097a8dd7a8d8b4d635afa837e88844cc37"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153744 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-mhxl4" event={"ID":"c62912da-e574-4877-9c1a-59e2a10c1d18","Type":"ContainerStarted","Data":"aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153760 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-mhxl4" event={"ID":"c62912da-e574-4877-9c1a-59e2a10c1d18","Type":"ContainerStarted","Data":"0295974055bea48df613818770c47872dc587a7d1f1c44c872d9ae4f645004f7"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153774 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153789 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153805 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"dcec7b0f39f51ff254e8c971d1938a64fa81a2d3979421a5cb2a500decda9ca3"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153822 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerStarted","Data":"7f74c24b21e0e98dfa2cad0d63178fab428b9c252845a89ae56a5419381c6a5c"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153837 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"32fa1ab5fb007de8f7a4bde1966442637e43e2c92aba19d69d8547df61db35d6"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153852 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153869 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"737f652182fd4bd2803099832c1c5c2b26df3a64a4e4ad4e16a4451b8d3b080a"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153886 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerStarted","Data":"7e0445ff27f2ebd8880ddce101fef6bbed394364d40f43f254bf6332cd4653ad"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153902 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.153917 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e8634ce928110631d79eecbf6209a89118a18aa244a713d7519a4e802b966c0d"} Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.498320 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.501823 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.516820 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.524776 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.524928 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.524967 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.525052 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.525113 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:27.525093988 +0000 UTC m=+23.136520812 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.525165 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:27.52515651 +0000 UTC m=+23.136583334 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.525232 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.525261 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:27.525251892 +0000 UTC m=+23.136678716 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.553038 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.567784 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.569959 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.625118 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.625442 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.625508 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625667 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625691 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625704 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625737 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625787 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625803 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625770 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:27.62575257 +0000 UTC m=+23.237179394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.625876 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:27.625855673 +0000 UTC m=+23.237282497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.641065 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.648693 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.717554 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.724385 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.725962 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.748133 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.751942 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.765696 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.776034 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.782693 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.788661 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.790243 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.796185 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.809400 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.823984 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.827213 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.835785 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.844597 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.863317 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.876589 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.878006 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.881720 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.882844 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.882922 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.882845 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.882981 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.883107 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:25 crc kubenswrapper[4921]: E0103 03:41:25.883214 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.892115 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.903388 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.908358 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.910736 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.920303 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.928477 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.942107 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.947256 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.957351 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.958264 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.971779 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.984718 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.984759 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.984807 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.996254 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:25 crc kubenswrapper[4921]: I0103 03:41:25.997640 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.020319 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.037869 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.054292 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.068200 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.080758 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.086311 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5"} Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.086385 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625"} Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.087619 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" exitCode=0 Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.087642 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.089857 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.090541 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36" exitCode=0 Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.090621 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36"} Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.101063 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.102228 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.105474 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerStarted","Data":"ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f"} Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.130058 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.163616 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.217723 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.218390 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.222032 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.247974 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.251249 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.257183 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.260662 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.262184 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.280898 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.305406 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.316366 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.318710 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.334574 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.358933 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.373162 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.394884 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.410407 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.425062 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.438558 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.457158 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.469710 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.496811 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.520856 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.538543 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.562310 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.579725 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.600684 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:26 crc kubenswrapper[4921]: I0103 03:41:26.614743 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:26Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111847 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111897 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111907 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111917 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111925 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.111934 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.113708 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d" exitCode=0 Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.113762 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.115865 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.128085 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.148571 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.171163 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.185907 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.199167 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.214027 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.230672 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.241980 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.250925 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.274424 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.312505 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.330868 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.347712 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.368044 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.382719 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.401415 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.429592 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.441121 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.462186 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.479946 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.492763 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.506620 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.521642 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.534729 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.548331 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.556712 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.556897 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.557016 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.557017 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:31.556968512 +0000 UTC m=+27.168395426 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.557088 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:31.557069285 +0000 UTC m=+27.168496109 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.557413 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.557564 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:31.557536297 +0000 UTC m=+27.168963161 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.557620 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.561220 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.627226 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.648414 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.659035 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.659126 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659374 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659420 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659444 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659519 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659548 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:31.659514724 +0000 UTC m=+27.270941588 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659555 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659593 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.659696 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:31.659670298 +0000 UTC m=+27.271097162 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.662683 4921 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.664955 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.664997 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.665015 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.665114 4921 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.672138 4921 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.672503 4921 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.674092 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.674117 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.674127 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.674139 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.674149 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.744536 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.750950 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.751067 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.751135 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.751217 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.751301 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.770170 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.774216 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.774329 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.774405 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.774486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.774544 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.788508 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.792710 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.792757 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.792769 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.792790 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.792803 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.804786 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.808464 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.808587 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.808654 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.808734 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.808800 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.825425 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:27Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.825945 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.827871 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.827968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.828028 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.828100 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.828166 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.883496 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.883559 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.883630 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.883657 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.883793 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:27 crc kubenswrapper[4921]: E0103 03:41:27.883956 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.930689 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.930781 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.930810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.930856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:27 crc kubenswrapper[4921]: I0103 03:41:27.930884 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:27Z","lastTransitionTime":"2026-01-03T03:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.034754 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.034796 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.034805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.034822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.034832 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.121741 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24" exitCode=0 Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.121890 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.137566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.137629 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.137643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.137669 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.137686 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.145584 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.168965 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.188684 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.205513 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.234213 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.250451 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.251121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.253557 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.253675 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.253699 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.256293 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.272305 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.286175 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.300936 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.315321 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.332661 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.356566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.356611 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.356625 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.356642 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.356656 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.357114 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.368584 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.382399 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:28Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.459414 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.459735 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.459838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.459943 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.460028 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.563103 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.563212 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.563228 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.563249 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.563280 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.666313 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.666367 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.666378 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.666400 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.666413 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.769928 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.770001 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.770021 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.770050 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.770069 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.873260 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.873313 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.873325 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.873340 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.873351 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.975687 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.975730 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.975739 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.975759 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:28 crc kubenswrapper[4921]: I0103 03:41:28.975771 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:28Z","lastTransitionTime":"2026-01-03T03:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.077742 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.077776 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.077785 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.077801 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.077811 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.130221 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.133601 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04" exitCode=0 Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.133675 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.150697 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.170061 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.180173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.180225 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.180250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.180324 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.180349 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.185654 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.209724 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.238254 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.265741 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.284237 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.284323 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.284342 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.284397 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.284416 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.287718 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.302868 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.318965 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.331965 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.351653 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.368876 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387309 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387531 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387589 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387610 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387636 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.387655 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.401399 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.491323 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.491381 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.491395 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.491420 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.491433 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.594945 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.594993 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.595005 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.595023 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.595035 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.698852 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.698926 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.698942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.698968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.698987 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.803247 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.803330 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.803347 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.803401 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.803422 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.884489 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.884552 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.884508 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:29 crc kubenswrapper[4921]: E0103 03:41:29.884785 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:29 crc kubenswrapper[4921]: E0103 03:41:29.884925 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:29 crc kubenswrapper[4921]: E0103 03:41:29.885043 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.910578 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.910649 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.910668 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.910699 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:29 crc kubenswrapper[4921]: I0103 03:41:29.910729 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:29Z","lastTransitionTime":"2026-01-03T03:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.014838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.014901 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.015047 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.015081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.015130 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.118709 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.118816 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.118831 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.118852 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.118866 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.143445 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e" exitCode=0 Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.143505 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.173380 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.206444 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.222901 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.222938 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.222950 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.222969 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.222980 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.225044 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.246626 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.260449 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.275006 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.301831 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.322090 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.326464 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.326498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.326507 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.326523 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.326534 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.341046 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.356959 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.370215 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.387440 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.404355 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.417967 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:30Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.429809 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.429861 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.429875 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.429897 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.429912 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.532967 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.534337 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.534349 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.534365 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.534375 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.637429 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.637478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.637488 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.637614 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.637636 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.740204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.740248 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.740288 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.740306 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.740319 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.843017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.843074 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.843091 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.843118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.843136 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.946021 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.946094 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.946112 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.946144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:30 crc kubenswrapper[4921]: I0103 03:41:30.946161 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:30Z","lastTransitionTime":"2026-01-03T03:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.049168 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.049214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.049228 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.049245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.049256 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.151785 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.151825 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.151836 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.151858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.151870 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.153871 4921 generic.go:334] "Generic (PLEG): container finished" podID="165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f" containerID="5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7" exitCode=0 Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.153905 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerDied","Data":"5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.177627 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.195549 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.211649 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.233482 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.259446 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.260486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.260536 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.260548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.260564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.260574 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.276546 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.289531 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.303948 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.315647 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.328564 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.340003 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.357624 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.363145 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.363180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.363190 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.363210 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.363222 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.372130 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.389399 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:31Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.466491 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.466549 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.466562 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.466604 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.466620 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.576448 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.576532 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.576554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.576606 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.576634 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.605945 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.606664 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.606615523 +0000 UTC m=+35.218042387 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.607541 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.607880 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.607700 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.608264 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.608244007 +0000 UTC m=+35.219670861 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.608106 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.608586 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.608568685 +0000 UTC m=+35.219995549 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.679573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.679629 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.679676 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.679720 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.679748 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.709716 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.710167 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710061 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710415 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710442 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710545 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.710518541 +0000 UTC m=+35.321945375 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710612 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710677 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710697 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.710799 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.710772968 +0000 UTC m=+35.322200012 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.783390 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.783453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.783520 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.783546 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.783564 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.882989 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.883052 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.883141 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.883226 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.883317 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:31 crc kubenswrapper[4921]: E0103 03:41:31.883379 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.889519 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.889566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.889582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.889607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.889626 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.993471 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.993542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.993557 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.993580 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:31 crc kubenswrapper[4921]: I0103 03:41:31.993594 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:31Z","lastTransitionTime":"2026-01-03T03:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.097023 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.097081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.097093 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.097110 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.097125 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.161617 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" event={"ID":"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f","Type":"ContainerStarted","Data":"edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.170316 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.170952 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.171017 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.175500 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.188342 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.196774 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199286 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199748 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199869 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.199887 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.202372 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.213428 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.222495 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.232311 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.244414 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.256408 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.272391 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.296570 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.303515 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.303596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.303623 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.303658 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.303687 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.315618 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.331417 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.343187 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.354221 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.364242 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.376728 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.388632 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.401748 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.405918 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.405968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.405980 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.406001 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.406014 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.414307 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.431031 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.448818 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.460254 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.470153 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.489882 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.509594 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.509648 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.509659 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.509681 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.509693 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.513393 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.525411 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.537153 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.553644 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:32Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.613022 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.613068 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.613082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.613099 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.613111 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.716262 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.716354 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.716372 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.716401 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.716421 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.820231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.820327 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.820343 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.820368 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.820384 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.923550 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.923631 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.923655 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.923688 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:32 crc kubenswrapper[4921]: I0103 03:41:32.923723 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:32Z","lastTransitionTime":"2026-01-03T03:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.027692 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.027781 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.027810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.027847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.027871 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.131037 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.131124 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.131151 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.131187 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.131208 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.175066 4921 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.234660 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.234734 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.234753 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.234782 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.234801 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.338499 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.338565 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.338575 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.338598 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.338612 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.441478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.441532 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.441551 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.441574 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.441594 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.544077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.544149 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.544168 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.544199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.544220 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.647535 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.647582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.647595 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.647616 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.647629 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.750748 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.750813 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.750828 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.750848 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.750860 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.853132 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.853178 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.853189 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.853205 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.853218 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.883824 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.883870 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.883846 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:33 crc kubenswrapper[4921]: E0103 03:41:33.884073 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:33 crc kubenswrapper[4921]: E0103 03:41:33.884232 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:33 crc kubenswrapper[4921]: E0103 03:41:33.884575 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.956621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.956732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.956763 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.956800 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:33 crc kubenswrapper[4921]: I0103 03:41:33.956822 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:33Z","lastTransitionTime":"2026-01-03T03:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.059214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.059322 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.059344 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.059380 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.059402 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.163097 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.163180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.163207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.163238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.163262 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.181556 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/0.log" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.185549 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4" exitCode=1 Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.185606 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.187007 4921 scope.go:117] "RemoveContainer" containerID="0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.206722 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.234815 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.251801 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.264736 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.266757 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.266821 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.266841 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.266874 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.266898 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.277960 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.295630 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.318443 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.337121 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.352995 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.367738 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.370101 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.370186 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.370203 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.370224 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.370237 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.383113 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.402790 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.439490 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:34Z\\\",\\\"message\\\":\\\"hift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:34.029969 6225 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030027 6225 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030088 6225 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030576 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:34.030595 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:34.030613 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:34.030619 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:34.030690 6225 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030924 6225 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.031523 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0103 03:41:34.031558 6225 factory.go:656] Stopping watch factory\\\\nI0103 03:41:34.031574 6225 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.461756 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.473121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.473169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.473180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.473199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.473210 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.576664 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.576720 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.576735 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.576755 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.576770 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.684002 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.684049 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.684060 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.684077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.684091 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.786801 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.786844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.786854 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.786871 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.786883 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.888692 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.888738 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.888750 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.888766 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.888777 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.900439 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.918103 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.936261 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.953932 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.965855 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.983612 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.991921 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.991968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.991985 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.992011 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.992029 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:34Z","lastTransitionTime":"2026-01-03T03:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:34 crc kubenswrapper[4921]: I0103 03:41:34.998172 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:34Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.011443 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.022384 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.037480 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.063820 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:34Z\\\",\\\"message\\\":\\\"hift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:34.029969 6225 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030027 6225 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030088 6225 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030576 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:34.030595 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:34.030613 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:34.030619 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:34.030690 6225 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030924 6225 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.031523 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0103 03:41:34.031558 6225 factory.go:656] Stopping watch factory\\\\nI0103 03:41:34.031574 6225 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.091468 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.094495 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.094531 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.094544 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.094566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.094581 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.110129 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.123445 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.190853 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/0.log" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.193604 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.193709 4921 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.196923 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.196962 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.196973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.196987 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.196998 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.206938 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.219901 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.235329 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.248354 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.260359 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.272689 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.288251 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.299729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.299768 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.299778 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.299796 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.299811 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.300455 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.312562 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.327524 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.347671 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:34Z\\\",\\\"message\\\":\\\"hift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:34.029969 6225 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030027 6225 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030088 6225 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030576 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:34.030595 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:34.030613 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:34.030619 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:34.030690 6225 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030924 6225 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.031523 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0103 03:41:34.031558 6225 factory.go:656] Stopping watch factory\\\\nI0103 03:41:34.031574 6225 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.364314 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.382567 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.400462 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:35Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.402098 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.402143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.402156 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.402176 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.402190 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.505557 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.505598 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.505608 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.505623 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.505636 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.607754 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.607808 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.607822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.607843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.607860 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.710762 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.710797 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.710807 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.710824 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.710835 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.814121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.814191 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.814211 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.814242 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.814264 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.882990 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.883054 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.883187 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:35 crc kubenswrapper[4921]: E0103 03:41:35.883306 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:35 crc kubenswrapper[4921]: E0103 03:41:35.883521 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:35 crc kubenswrapper[4921]: E0103 03:41:35.883626 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.919540 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.919637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.919665 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.919702 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:35 crc kubenswrapper[4921]: I0103 03:41:35.919744 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:35Z","lastTransitionTime":"2026-01-03T03:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.024361 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.024449 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.024477 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.024510 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.024533 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.127213 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.127263 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.127296 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.127315 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.127330 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.199764 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/1.log" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.200438 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/0.log" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.205226 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec" exitCode=1 Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.205318 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.205403 4921 scope.go:117] "RemoveContainer" containerID="0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.206372 4921 scope.go:117] "RemoveContainer" containerID="f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec" Jan 03 03:41:36 crc kubenswrapper[4921]: E0103 03:41:36.207371 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.229770 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.231384 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.231499 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.231562 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.231595 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.231643 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.246868 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.266964 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.289283 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.307502 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.334225 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.335692 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.335766 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.335787 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.335820 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.335844 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.369121 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b2055636fa3a25a8ef578aabd8589a421fbe36c05d992b04b32d7f5e351ffa4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:34Z\\\",\\\"message\\\":\\\"hift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:34.029969 6225 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030027 6225 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030088 6225 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030576 6225 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:34.030595 6225 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:34.030613 6225 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:34.030619 6225 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:34.030690 6225 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.030924 6225 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0103 03:41:34.031523 6225 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0103 03:41:34.031558 6225 factory.go:656] Stopping watch factory\\\\nI0103 03:41:34.031574 6225 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.393180 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.411122 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.433510 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.438813 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.438867 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.438883 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.438908 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.438927 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.452573 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.469514 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.486641 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.505989 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:36Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.542388 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.542447 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.542465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.542495 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.542678 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.645627 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.645973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.646090 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.646231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.646373 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.749996 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.750043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.750059 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.750081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.750096 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.852963 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.853014 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.853026 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.853046 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.853059 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.955710 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.955765 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.955823 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.955847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:36 crc kubenswrapper[4921]: I0103 03:41:36.955860 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:36Z","lastTransitionTime":"2026-01-03T03:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.059227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.059288 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.059300 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.059319 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.059331 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.169088 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.169146 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.169160 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.169180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.169195 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.212003 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/1.log" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.220091 4921 scope.go:117] "RemoveContainer" containerID="f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.220466 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.238412 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.258872 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.271872 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.271924 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.271941 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.271968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.271987 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.288170 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.315980 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.336680 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.358006 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.374816 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.374880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.374896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.374919 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.374934 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.379051 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.399078 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.415783 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.438084 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.462679 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.478322 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.478383 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.478396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.478417 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.478431 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.483398 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.497851 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.503434 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.523682 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.541204 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.558650 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.573360 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.581576 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.581623 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.581643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.581690 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.581712 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.588408 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.607922 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.627001 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.645859 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.662810 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.679573 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.686748 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.686823 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.686849 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.686880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.686905 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.696503 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p"] Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.697243 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.699733 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.700380 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.702940 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.718573 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.732289 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.753002 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.774997 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.775066 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.775092 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q8cq\" (UniqueName: \"kubernetes.io/projected/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-kube-api-access-8q8cq\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.775366 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.789718 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.790337 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.790372 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.790386 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.790406 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.790421 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.820770 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.835678 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.854192 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.874401 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.877362 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.877640 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.877748 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q8cq\" (UniqueName: \"kubernetes.io/projected/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-kube-api-access-8q8cq\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.877912 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.878331 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.878331 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.884110 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.884170 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.884243 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.884373 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.884110 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.884533 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.892261 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.893810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.893866 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.893886 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.893915 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.893941 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.903907 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.908498 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q8cq\" (UniqueName: \"kubernetes.io/projected/bf98706c-cd2f-4ab8-98aa-ed60ce7dd404-kube-api-access-8q8cq\") pod \"ovnkube-control-plane-749d76644c-cnc2p\" (UID: \"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.924031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.924223 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.924423 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.924573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.924696 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.927492 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.942211 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.946099 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.947921 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.947983 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.948007 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.948040 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.948060 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.968728 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.970636 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.976503 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.976572 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.976586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.976613 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.976629 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:37Z","lastTransitionTime":"2026-01-03T03:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:37 crc kubenswrapper[4921]: I0103 03:41:37.985213 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:37 crc kubenswrapper[4921]: E0103 03:41:37.996407 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:37Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.003226 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.003304 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.003326 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.003354 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.003376 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.007858 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.015134 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.031312 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.031954 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.037798 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.037843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.037856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.037879 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.037896 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.053019 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.059446 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.059572 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.063627 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.063654 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.063664 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.063682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.063697 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.080834 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.095163 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.108467 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.168440 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.168500 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.168516 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.168536 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.168559 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.221958 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" event={"ID":"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404","Type":"ContainerStarted","Data":"8dd5a351ba13029070b219433885104022ce65519f3f528aa0208c479c782f8e"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.271573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.271625 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.271634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.271653 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.271665 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.374446 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.374489 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.374499 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.374517 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.374530 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.453756 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vsp88"] Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.454881 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.455036 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.471071 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.478144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.478204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.478224 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.478253 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.478303 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.486878 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.509070 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.530764 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.548959 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.567912 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.582250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.582341 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.582364 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.582394 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.582418 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.585512 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.585606 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snpwl\" (UniqueName: \"kubernetes.io/projected/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-kube-api-access-snpwl\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.590010 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.609896 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.632519 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.651557 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.674672 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.684844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.684917 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.684930 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.684974 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.684993 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.687124 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snpwl\" (UniqueName: \"kubernetes.io/projected/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-kube-api-access-snpwl\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.687402 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.687749 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:38 crc kubenswrapper[4921]: E0103 03:41:38.687833 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:39.187810391 +0000 UTC m=+34.799237225 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.692226 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.722167 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.722737 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snpwl\" (UniqueName: \"kubernetes.io/projected/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-kube-api-access-snpwl\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.754002 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.774911 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.788607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.788683 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.788704 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.788732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.788755 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.794988 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:38Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.899496 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.899548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.899565 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.899589 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:38 crc kubenswrapper[4921]: I0103 03:41:38.899609 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:38Z","lastTransitionTime":"2026-01-03T03:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.004573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.004639 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.004656 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.004682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.004700 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.108411 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.108471 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.108486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.108510 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.108527 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.200773 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.201199 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.201358 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:40.201326041 +0000 UTC m=+35.812752875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.211242 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.211305 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.211320 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.211342 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.211357 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.227995 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" event={"ID":"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404","Type":"ContainerStarted","Data":"4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.314723 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.314788 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.314810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.314839 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.314866 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.417988 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.418046 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.418057 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.418125 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.418154 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.521752 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.521810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.521822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.521844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.521858 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.625337 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.625383 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.625392 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.625410 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.625421 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.707407 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.707562 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.707660 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.707683 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:55.707643101 +0000 UTC m=+51.319069935 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.707725 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:55.707713412 +0000 UTC m=+51.319140246 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.707842 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.708012 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.708068 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:55.708053471 +0000 UTC m=+51.319480295 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.728169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.728194 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.728203 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.728217 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.728227 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.809373 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.809486 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809679 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809708 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809726 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809767 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809819 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:55.809797022 +0000 UTC m=+51.421223856 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809847 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.809888 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.810038 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:41:55.809994268 +0000 UTC m=+51.421421272 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.831722 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.831803 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.831826 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.831855 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.831881 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.883883 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.883943 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.883914 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.884072 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.884024 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.884192 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.884419 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:39 crc kubenswrapper[4921]: E0103 03:41:39.884589 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.935969 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.936025 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.936043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.936073 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:39 crc kubenswrapper[4921]: I0103 03:41:39.936092 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:39Z","lastTransitionTime":"2026-01-03T03:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.038940 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.038989 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.039001 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.039020 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.039030 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.142749 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.142846 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.142877 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.142921 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.142950 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.214943 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:40 crc kubenswrapper[4921]: E0103 03:41:40.215127 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:40 crc kubenswrapper[4921]: E0103 03:41:40.215224 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:42.215197923 +0000 UTC m=+37.826624747 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.232204 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" event={"ID":"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404","Type":"ContainerStarted","Data":"33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.246262 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.246387 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.246427 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.246457 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.246477 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.247576 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.265951 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.278015 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.293930 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.306374 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.318382 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.337635 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.349153 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.349194 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.349204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.349225 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.349239 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.355413 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.370629 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.383700 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.399006 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.419030 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.432185 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.452503 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.452548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.452559 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.452577 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.452589 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.455385 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.485637 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.511327 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:40Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.555124 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.555201 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.555214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.555235 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.555248 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.660824 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.660937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.660952 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.660972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.660989 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.763442 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.763478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.763515 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.763531 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.763541 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.866621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.866720 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.866751 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.866790 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.866810 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.970047 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.970092 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.970104 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.970122 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:40 crc kubenswrapper[4921]: I0103 03:41:40.970135 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:40Z","lastTransitionTime":"2026-01-03T03:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.073262 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.073333 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.073346 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.073367 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.073380 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.175612 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.175657 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.175673 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.175693 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.175705 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.278543 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.278624 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.278648 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.278679 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.278701 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.381180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.381224 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.381235 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.381250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.381260 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.484221 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.484279 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.484288 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.484305 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.484316 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.586596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.586654 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.586668 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.586687 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.586701 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.689169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.689213 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.689221 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.689239 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.689249 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.791297 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.791383 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.791395 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.791414 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.791426 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.883020 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.883108 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:41 crc kubenswrapper[4921]: E0103 03:41:41.883184 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.883019 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.883045 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:41 crc kubenswrapper[4921]: E0103 03:41:41.883246 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:41 crc kubenswrapper[4921]: E0103 03:41:41.883437 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:41 crc kubenswrapper[4921]: E0103 03:41:41.883571 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.894036 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.894099 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.894118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.894147 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.894164 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.997979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.998029 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.998038 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.998055 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:41 crc kubenswrapper[4921]: I0103 03:41:41.998065 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:41Z","lastTransitionTime":"2026-01-03T03:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.101238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.101308 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.101318 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.101337 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.101349 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.206006 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.206068 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.206078 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.206095 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.206115 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.237008 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:42 crc kubenswrapper[4921]: E0103 03:41:42.237183 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:42 crc kubenswrapper[4921]: E0103 03:41:42.237317 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:46.237291726 +0000 UTC m=+41.848718590 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.309547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.309633 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.309657 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.309695 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.309720 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.413524 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.413586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.413612 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.413645 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.413671 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.516843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.516925 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.516947 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.516976 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.516996 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.620732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.620797 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.620814 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.620843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.620861 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.724423 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.724475 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.724493 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.724518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.724538 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.827379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.827453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.827476 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.827511 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.827533 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.930818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.930890 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.930931 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.930970 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:42 crc kubenswrapper[4921]: I0103 03:41:42.930994 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:42Z","lastTransitionTime":"2026-01-03T03:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.034917 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.035006 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.035031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.035064 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.035085 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.141168 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.141244 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.141264 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.141359 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.141386 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.244572 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.244634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.244656 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.244682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.244703 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.348654 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.348711 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.348724 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.348747 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.348765 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.452491 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.452539 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.452552 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.452578 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.452591 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.555826 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.555884 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.555903 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.555936 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.555957 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.659494 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.659552 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.659570 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.659593 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.659607 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.763028 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.763104 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.763135 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.763174 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.763197 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.866004 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.866054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.866070 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.866096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.866113 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.883600 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:43 crc kubenswrapper[4921]: E0103 03:41:43.883802 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.884372 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:43 crc kubenswrapper[4921]: E0103 03:41:43.884544 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.884572 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:43 crc kubenswrapper[4921]: E0103 03:41:43.884770 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.884885 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:43 crc kubenswrapper[4921]: E0103 03:41:43.885133 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.969760 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.969856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.969877 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.969939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:43 crc kubenswrapper[4921]: I0103 03:41:43.969958 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:43Z","lastTransitionTime":"2026-01-03T03:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.074088 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.074153 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.074201 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.074231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.074249 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.177308 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.177355 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.177364 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.177382 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.177394 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.280344 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.280419 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.280446 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.280480 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.280503 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.384336 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.384416 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.384442 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.384485 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.384507 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.487530 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.487608 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.487628 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.487659 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.487685 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.591065 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.591138 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.591172 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.591207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.591231 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.694351 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.694415 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.694431 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.694453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.694467 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.797183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.797245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.797263 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.797335 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.797360 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.900677 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.900761 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.900786 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.900818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.900841 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:44Z","lastTransitionTime":"2026-01-03T03:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.905067 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:44Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.930974 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:44Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.954878 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:44Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.972844 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:44Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:44 crc kubenswrapper[4921]: I0103 03:41:44.990476 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:44Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.004339 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.004407 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.004430 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.004461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.004484 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.016567 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.038133 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.058338 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.080146 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.107896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.107964 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.107986 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.108018 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.108039 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.113207 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.136098 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.158511 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.178730 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.194741 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.211661 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.211742 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.211766 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.211795 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.211815 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.216191 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.231314 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:45Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.315535 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.315589 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.315601 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.315624 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.315637 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.419756 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.419864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.419890 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.419925 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.419947 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.524716 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.524795 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.524815 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.524846 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.524873 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.628898 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.628972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.628992 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.629019 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.629039 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.732650 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.732725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.732743 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.732778 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.732797 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.836330 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.836405 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.836426 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.836452 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.836471 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.882830 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.882930 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:45 crc kubenswrapper[4921]: E0103 03:41:45.883012 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.883038 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.883066 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:45 crc kubenswrapper[4921]: E0103 03:41:45.883231 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:45 crc kubenswrapper[4921]: E0103 03:41:45.883393 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:45 crc kubenswrapper[4921]: E0103 03:41:45.883533 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.947248 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.947685 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.947856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.947994 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:45 crc kubenswrapper[4921]: I0103 03:41:45.948182 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:45Z","lastTransitionTime":"2026-01-03T03:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.052395 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.052456 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.052466 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.052492 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.052504 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.155556 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.156065 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.156225 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.156465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.156619 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.259769 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.259835 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.259855 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.259882 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.259901 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.288620 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:46 crc kubenswrapper[4921]: E0103 03:41:46.288909 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:46 crc kubenswrapper[4921]: E0103 03:41:46.289051 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:41:54.289022331 +0000 UTC m=+49.900449165 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.363384 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.363468 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.363490 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.363526 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.363547 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.467129 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.467214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.467234 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.467265 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.467328 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.570971 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.571064 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.571082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.571107 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.571175 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.674109 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.674173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.674191 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.674218 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.674238 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.777396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.777472 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.777485 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.777504 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.777518 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.881860 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.881932 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.881950 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.881979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.881997 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.985784 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.985858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.985883 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.985915 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:46 crc kubenswrapper[4921]: I0103 03:41:46.985935 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:46Z","lastTransitionTime":"2026-01-03T03:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.090143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.090245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.090259 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.090299 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.090315 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.193297 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.193361 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.193371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.193389 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.193403 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.297526 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.297611 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.297632 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.297705 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.297728 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.342413 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.344073 4921 scope.go:117] "RemoveContainer" containerID="f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.401965 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.402392 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.402405 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.402424 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.402439 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.506396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.506482 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.506511 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.506547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.506573 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.609859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.609896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.609907 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.609926 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.609940 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.724639 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.724710 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.724736 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.724772 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.724792 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.828356 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.828400 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.828411 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.828430 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.828442 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.883153 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.883262 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:47 crc kubenswrapper[4921]: E0103 03:41:47.883323 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.883345 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.883368 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:47 crc kubenswrapper[4921]: E0103 03:41:47.883488 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:47 crc kubenswrapper[4921]: E0103 03:41:47.883562 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:47 crc kubenswrapper[4921]: E0103 03:41:47.883609 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.931612 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.931652 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.931663 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.931679 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:47 crc kubenswrapper[4921]: I0103 03:41:47.931701 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:47Z","lastTransitionTime":"2026-01-03T03:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.034934 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.034969 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.034979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.035029 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.035041 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.108986 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.109044 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.109060 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.109083 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.109097 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.126221 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.130199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.130248 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.130258 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.130294 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.130308 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.146306 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.150953 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.151022 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.151041 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.151071 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.151091 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.167553 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.173379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.173434 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.173446 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.173468 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.173480 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.188420 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.193473 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.193533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.193549 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.193571 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.193585 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.208586 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: E0103 03:41:48.208771 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.211095 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.211149 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.211160 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.211183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.211198 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.264777 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/1.log" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.267581 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.268056 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.289120 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.306322 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.314682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.314744 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.314953 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.314982 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.315005 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.321918 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.347950 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.371436 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.392351 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.408485 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.417401 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.417461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.417475 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.417497 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.417511 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.421745 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.432413 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.453469 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.487001 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.501474 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.520670 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.520981 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.521039 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.521050 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.521068 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.521077 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.540340 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.559378 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.580713 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:48Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.624476 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.624517 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.624533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.624552 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.624565 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.727310 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.727347 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.727356 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.727371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.727384 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.831828 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.831902 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.831920 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.831947 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.831967 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.935995 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.936076 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.936096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.936120 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:48 crc kubenswrapper[4921]: I0103 03:41:48.936138 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:48Z","lastTransitionTime":"2026-01-03T03:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.038099 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.038134 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.038143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.038157 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.038167 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.141574 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.141634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.141649 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.141668 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.141681 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.244909 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.245017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.245047 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.245084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.245112 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.274739 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/2.log" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.275651 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/1.log" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.279576 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" exitCode=1 Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.279632 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.279690 4921 scope.go:117] "RemoveContainer" containerID="f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.280583 4921 scope.go:117] "RemoveContainer" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" Jan 03 03:41:49 crc kubenswrapper[4921]: E0103 03:41:49.280813 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.303070 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.317814 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.333943 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.347195 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.348126 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.348180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.348194 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.348215 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.348230 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.366771 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.385068 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.401233 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.431625 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9f2586e61011ac88c31161889190e29ff4396fddc2197a26b0cd2025f88d5ec\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:35Z\\\",\\\"message\\\":\\\"pping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178580 6350 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.178890 6350 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0103 03:41:35.178923 6350 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0103 03:41:35.179068 6350 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:35.179084 6350 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:35.179108 6350 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0103 03:41:35.179151 6350 factory.go:656] Stopping watch factory\\\\nI0103 03:41:35.179171 6350 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0103 03:41:35.179181 6350 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:35.179191 6350 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.449158 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.451153 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.451215 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.451237 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.451303 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.451325 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.466421 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.481162 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.500978 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.519010 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.535842 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.552545 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.554741 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.554815 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.554832 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.554883 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.554901 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.566132 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:49Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.657889 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.657959 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.657977 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.658009 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.658033 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.761682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.761759 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.761783 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.761817 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.761837 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.865559 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.865628 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.865647 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.865675 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.865695 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.883143 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.883160 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.883182 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.883350 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:49 crc kubenswrapper[4921]: E0103 03:41:49.883562 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:49 crc kubenswrapper[4921]: E0103 03:41:49.883706 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:49 crc kubenswrapper[4921]: E0103 03:41:49.883853 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:49 crc kubenswrapper[4921]: E0103 03:41:49.884013 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.969174 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.969259 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.969334 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.969370 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:49 crc kubenswrapper[4921]: I0103 03:41:49.969396 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:49Z","lastTransitionTime":"2026-01-03T03:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.072131 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.072211 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.072231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.072264 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.072358 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.175963 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.176053 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.176080 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.176121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.176148 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.279453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.279516 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.279527 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.279549 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.279562 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.286021 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/2.log" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.297315 4921 scope.go:117] "RemoveContainer" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" Jan 03 03:41:50 crc kubenswrapper[4921]: E0103 03:41:50.297651 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.314442 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.331591 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.344479 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.359548 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.377873 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.382512 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.382696 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.382724 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.382753 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.382776 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.401141 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.420228 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.434599 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.460335 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.484980 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.486538 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.486607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.486627 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.486651 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.486666 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.502777 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.519208 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.534980 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.551057 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.565565 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.583338 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:50Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.589157 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.589207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.589226 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.589252 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.589291 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.692448 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.692500 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.692513 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.692534 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.692547 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.795883 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.795937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.795953 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.795976 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.795992 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.899060 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.899107 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.899118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.899136 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:50 crc kubenswrapper[4921]: I0103 03:41:50.899153 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:50Z","lastTransitionTime":"2026-01-03T03:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.003083 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.003156 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.003176 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.003206 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.003226 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.106480 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.106533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.106543 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.106584 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.106596 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.209357 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.209399 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.209411 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.209431 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.209448 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.312861 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.312930 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.312947 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.312970 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.312987 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.416339 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.416400 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.416412 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.416485 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.416502 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.520000 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.520113 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.520213 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.520251 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.520384 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.624355 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.624410 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.624427 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.624453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.624470 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.727911 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.727987 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.728008 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.728042 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.728068 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.831414 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.831507 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.831530 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.831559 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.831577 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.883744 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.883784 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.883784 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:51 crc kubenswrapper[4921]: E0103 03:41:51.883978 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.884073 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:51 crc kubenswrapper[4921]: E0103 03:41:51.884169 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:51 crc kubenswrapper[4921]: E0103 03:41:51.884316 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:51 crc kubenswrapper[4921]: E0103 03:41:51.884433 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.934931 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.934993 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.935010 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.935035 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:51 crc kubenswrapper[4921]: I0103 03:41:51.935052 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:51Z","lastTransitionTime":"2026-01-03T03:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.038550 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.038651 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.038711 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.038736 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.038757 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.141453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.141537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.141564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.141590 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.141609 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.244660 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.244725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.244744 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.244770 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.244789 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.347354 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.347400 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.347410 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.347428 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.347440 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.450916 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.450975 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.450989 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.451015 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.451031 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.554070 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.554158 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.554181 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.554214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.554236 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.657785 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.657856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.657869 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.657887 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.657899 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.760641 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.760684 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.760695 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.760715 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.760727 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.863512 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.863571 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.863581 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.863598 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.863610 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.967461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.967535 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.967554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.967582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:52 crc kubenswrapper[4921]: I0103 03:41:52.967602 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:52Z","lastTransitionTime":"2026-01-03T03:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.070933 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.071010 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.071028 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.071061 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.071084 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.174481 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.174558 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.174578 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.174608 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.174628 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.276783 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.276833 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.276885 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.276908 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.276921 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.379427 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.379490 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.379507 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.379530 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.379548 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.482812 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.482858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.482887 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.482904 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.482914 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.586910 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.586973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.586990 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.587018 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.587036 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.690058 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.690142 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.690162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.690222 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.690244 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.794631 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.794707 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.794725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.794761 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.794780 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.883466 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.883508 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.883516 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.883634 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:53 crc kubenswrapper[4921]: E0103 03:41:53.883723 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:53 crc kubenswrapper[4921]: E0103 03:41:53.883901 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:53 crc kubenswrapper[4921]: E0103 03:41:53.884027 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:53 crc kubenswrapper[4921]: E0103 03:41:53.884327 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.898807 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.898884 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.898910 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.898939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:53 crc kubenswrapper[4921]: I0103 03:41:53.898957 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:53Z","lastTransitionTime":"2026-01-03T03:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.002308 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.002390 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.002409 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.002437 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.002464 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.106095 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.106161 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.106175 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.106202 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.106219 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.209620 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.209684 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.209698 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.209751 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.209773 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.289866 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:54 crc kubenswrapper[4921]: E0103 03:41:54.290088 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:54 crc kubenswrapper[4921]: E0103 03:41:54.290183 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:42:10.290153096 +0000 UTC m=+65.901579960 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.315436 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.315488 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.315498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.315518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.315530 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.420393 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.420455 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.420466 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.420486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.420498 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.523911 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.523967 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.523978 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.523998 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.524014 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.630726 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.630799 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.630819 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.630845 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.630863 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.734741 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.734812 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.734837 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.734870 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.734893 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.838894 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.838989 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.839017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.839052 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.839075 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.905533 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.918786 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.933324 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.941201 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.941227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.941235 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.941253 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.941263 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:54Z","lastTransitionTime":"2026-01-03T03:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.947829 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.962960 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:54 crc kubenswrapper[4921]: I0103 03:41:54.982081 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.001589 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:54Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.017345 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.029411 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.041599 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.046190 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.046250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.046287 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.046319 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.046334 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.056619 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.077976 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.092670 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.109147 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.121649 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.133034 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:55Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.149420 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.149472 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.149486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.149510 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.149524 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.252239 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.252310 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.252325 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.252345 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.252364 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.355383 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.355453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.355475 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.355504 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.355524 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.459247 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.459519 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.459545 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.459583 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.459604 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.562629 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.562704 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.562729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.562756 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.562777 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.666294 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.666645 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.666657 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.666673 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.666684 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.709135 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.709312 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.709415 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.709479 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:42:27.709444038 +0000 UTC m=+83.320870892 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.709585 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.709631 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.709661 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:42:27.709643303 +0000 UTC m=+83.321070157 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.709802 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:42:27.709761857 +0000 UTC m=+83.321188731 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.769934 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.770605 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.770663 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.770688 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.770702 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.810082 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.810233 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810417 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810467 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810483 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810520 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810560 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810570 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:42:27.810545289 +0000 UTC m=+83.421972113 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810582 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.810658 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:42:27.810636012 +0000 UTC m=+83.422062876 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.874534 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.874591 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.874601 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.874619 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.874632 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.883144 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.883175 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.883190 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.883165 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.883382 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.883544 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.883683 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:55 crc kubenswrapper[4921]: E0103 03:41:55.883851 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.978759 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.978816 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.978825 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.978848 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:55 crc kubenswrapper[4921]: I0103 03:41:55.978858 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:55Z","lastTransitionTime":"2026-01-03T03:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.082373 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.082469 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.082498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.082533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.082558 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.185779 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.185857 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.185875 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.185901 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.185919 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.288700 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.288744 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.288752 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.288768 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.288778 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.390942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.391019 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.391041 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.391076 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.391114 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.493736 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.493819 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.493845 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.493880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.493903 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.597010 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.597071 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.597091 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.597118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.597135 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.701086 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.701159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.701177 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.701208 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.701227 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.804521 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.804665 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.804685 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.804717 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.804741 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.908849 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.908930 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.908951 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.908981 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:56 crc kubenswrapper[4921]: I0103 03:41:56.909007 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:56Z","lastTransitionTime":"2026-01-03T03:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.012507 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.012590 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.012609 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.012676 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.012702 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.116118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.116191 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.116214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.116245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.116263 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.220415 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.220487 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.220505 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.220533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.220556 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.323599 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.323655 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.323669 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.323690 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.323706 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.426859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.426932 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.426950 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.426979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.426999 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.530463 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.530516 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.530525 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.530545 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.530557 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.633500 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.633580 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.633601 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.633633 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.633689 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.736537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.736579 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.736591 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.736610 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.736624 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.840154 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.840223 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.840242 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.840315 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.840339 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.883396 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.883445 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.883445 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.883501 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:57 crc kubenswrapper[4921]: E0103 03:41:57.883613 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:57 crc kubenswrapper[4921]: E0103 03:41:57.883756 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:57 crc kubenswrapper[4921]: E0103 03:41:57.883867 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:57 crc kubenswrapper[4921]: E0103 03:41:57.884060 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.943766 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.943821 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.943842 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.943869 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:57 crc kubenswrapper[4921]: I0103 03:41:57.943891 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:57Z","lastTransitionTime":"2026-01-03T03:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.047627 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.047771 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.047788 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.047814 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.047836 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.150947 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.150997 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.151008 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.151026 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.151038 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.253767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.253864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.253890 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.253925 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.253949 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.265024 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.265094 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.265113 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.265142 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.265168 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.288166 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.296344 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.296437 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.296466 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.296506 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.296531 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.319972 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.324413 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.324457 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.324474 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.324497 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.324511 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.345237 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.351129 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.351170 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.351183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.351205 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.351220 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.370387 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.376047 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.376098 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.376113 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.376139 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.376156 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.395488 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: E0103 03:41:58.395748 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.398223 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.398312 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.398335 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.398367 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.398390 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.502549 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.502617 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.502637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.502669 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.502691 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.606378 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.606455 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.606478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.606509 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.606532 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.709648 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.709703 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.709729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.709764 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.709788 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.745914 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.764221 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.769332 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.786132 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.804453 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.817833 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.817914 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.817933 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.817965 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.817986 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.826596 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.872368 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.909653 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.920579 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.920649 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.920670 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.920697 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.920716 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:58Z","lastTransitionTime":"2026-01-03T03:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.933070 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.950028 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.969494 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:58 crc kubenswrapper[4921]: I0103 03:41:58.990988 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:58Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.008409 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.023168 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.024214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.024314 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.024334 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.024371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.024391 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.037335 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.055307 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.070693 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.086176 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:41:59Z is after 2025-08-24T17:21:41Z" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.128418 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.128477 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.128495 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.128524 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.128544 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.231879 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.231926 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.231937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.231954 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.231967 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.334357 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.334438 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.334460 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.334493 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.334511 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.437784 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.437833 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.437846 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.437866 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.437882 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.540668 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.540726 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.540740 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.540760 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.540776 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.643621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.643676 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.643694 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.643719 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.643739 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.747553 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.747619 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.747637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.747664 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.747686 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.851090 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.851154 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.851168 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.851191 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.851210 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.883426 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.883501 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.883517 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:41:59 crc kubenswrapper[4921]: E0103 03:41:59.883630 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.883446 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:41:59 crc kubenswrapper[4921]: E0103 03:41:59.883923 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:41:59 crc kubenswrapper[4921]: E0103 03:41:59.883979 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:41:59 crc kubenswrapper[4921]: E0103 03:41:59.884094 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.954470 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.954578 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.954613 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.954652 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:41:59 crc kubenswrapper[4921]: I0103 03:41:59.954682 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:41:59Z","lastTransitionTime":"2026-01-03T03:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.057827 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.057887 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.057900 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.057923 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.057939 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.160985 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.161058 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.161081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.161112 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.161134 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.264357 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.264398 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.264408 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.264426 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.264438 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.369103 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.369169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.369190 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.369222 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.369245 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.473162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.473229 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.473250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.473347 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.473368 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.576417 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.576498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.576518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.576552 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.576573 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.680568 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.680644 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.680662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.680690 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.680713 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.783919 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.783998 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.784021 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.784050 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.784071 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.884832 4921 scope.go:117] "RemoveContainer" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" Jan 03 03:42:00 crc kubenswrapper[4921]: E0103 03:42:00.885185 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.886849 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.886893 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.886909 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.886930 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.886944 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.990635 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.990705 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.990725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.990754 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:00 crc kubenswrapper[4921]: I0103 03:42:00.990773 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:00Z","lastTransitionTime":"2026-01-03T03:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.093988 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.094049 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.094067 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.094085 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.094099 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.198502 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.198551 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.198565 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.198584 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.198596 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.301975 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.302038 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.302056 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.302082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.302102 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.406355 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.406442 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.406461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.406489 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.406510 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.510490 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.510561 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.510579 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.510609 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.510630 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.616682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.616772 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.616790 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.616827 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.616845 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.720687 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.720737 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.720751 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.720773 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.720784 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.825607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.825670 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.825682 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.825703 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.825720 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.883651 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.883708 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.883728 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.883805 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:01 crc kubenswrapper[4921]: E0103 03:42:01.883940 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:01 crc kubenswrapper[4921]: E0103 03:42:01.884101 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:01 crc kubenswrapper[4921]: E0103 03:42:01.884287 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:01 crc kubenswrapper[4921]: E0103 03:42:01.884416 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.929137 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.929202 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.929221 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.929252 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:01 crc kubenswrapper[4921]: I0103 03:42:01.929302 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:01Z","lastTransitionTime":"2026-01-03T03:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.032895 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.032973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.033000 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.033034 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.033058 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.135949 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.136010 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.136034 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.136062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.136083 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.239035 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.239110 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.239133 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.239164 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.239186 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.342008 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.342062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.342081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.342108 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.342126 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.445487 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.446026 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.446182 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.446436 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.446620 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.549814 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.549866 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.549880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.549897 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.549913 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.652758 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.652798 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.652810 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.652828 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.652840 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.756931 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.757513 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.757679 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.757819 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.757961 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.861524 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.861582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.861596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.861618 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.861628 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.965354 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.966429 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.966652 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.966836 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:02 crc kubenswrapper[4921]: I0103 03:42:02.966970 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:02Z","lastTransitionTime":"2026-01-03T03:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.070419 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.070504 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.070528 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.070564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.070589 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.174380 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.174433 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.174451 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.174479 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.174503 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.277777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.277889 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.277910 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.277938 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.277958 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.380673 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.381520 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.381688 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.381859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.382072 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.485106 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.485675 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.485851 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.486012 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.486144 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.589844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.589927 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.589958 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.589992 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.590014 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.693474 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.693537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.693550 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.693573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.693585 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.796591 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.796650 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.796667 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.796693 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.796713 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.883639 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:03 crc kubenswrapper[4921]: E0103 03:42:03.883884 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.883956 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.884041 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.884197 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:03 crc kubenswrapper[4921]: E0103 03:42:03.884258 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:03 crc kubenswrapper[4921]: E0103 03:42:03.884402 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:03 crc kubenswrapper[4921]: E0103 03:42:03.884603 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.900165 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.900227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.900247 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.900313 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:03 crc kubenswrapper[4921]: I0103 03:42:03.900344 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:03Z","lastTransitionTime":"2026-01-03T03:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.004200 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.004488 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.004518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.004550 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.004573 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.108872 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.108931 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.108949 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.108976 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.108994 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.212634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.212722 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.212743 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.212776 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.212797 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.316569 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.316650 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.316662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.316690 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.316711 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.420981 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.421045 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.421062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.421089 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.421108 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.523492 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.523551 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.523569 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.523593 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.523612 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.627883 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.627962 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.627979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.628008 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.628027 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.730861 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.730938 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.730958 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.730992 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.731142 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.834686 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.834744 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.834760 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.834783 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.834799 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.901889 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:04Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.922365 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:04Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.939534 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.939586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.939600 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.939620 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.939632 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:04Z","lastTransitionTime":"2026-01-03T03:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.944982 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:04Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.973485 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:04Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:04 crc kubenswrapper[4921]: I0103 03:42:04.989129 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:04Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.005042 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.019432 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.033917 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.043805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.043842 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.043854 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.043875 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.043887 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.046792 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.062599 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.091180 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.107250 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.120956 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.133603 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.145146 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.153487 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.153542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.153554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.153576 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.153589 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.158662 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.175437 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:05Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.256734 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.256805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.256828 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.256867 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.256889 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.361441 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.361546 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.361606 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.361641 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.361701 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.465990 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.466518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.466734 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.466974 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.467172 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.571317 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.571405 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.571425 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.571455 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.571477 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.674666 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.674735 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.674755 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.674787 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.674842 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.778358 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.778442 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.778468 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.778506 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.778530 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882786 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882844 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882805 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882790 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882971 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.882990 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.883026 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.883048 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.883015 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:05 crc kubenswrapper[4921]: E0103 03:42:05.883196 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:05 crc kubenswrapper[4921]: E0103 03:42:05.883325 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:05 crc kubenswrapper[4921]: E0103 03:42:05.883388 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:05 crc kubenswrapper[4921]: E0103 03:42:05.883501 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.987266 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.987367 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.987385 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.987414 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:05 crc kubenswrapper[4921]: I0103 03:42:05.987432 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:05Z","lastTransitionTime":"2026-01-03T03:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.091362 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.091421 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.091439 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.091464 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.091482 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.194939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.195503 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.195670 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.195842 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.196032 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.299572 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.299923 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.300084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.300237 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.300423 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.419095 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.419190 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.419212 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.419241 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.419260 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.522870 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.522956 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.522980 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.523016 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.523043 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.630776 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.630859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.630884 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.630917 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.630940 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.734465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.734528 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.734589 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.734616 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.734634 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.855051 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.855084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.855094 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.855109 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.855121 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.957767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.957847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.957857 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.957880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:06 crc kubenswrapper[4921]: I0103 03:42:06.957894 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:06Z","lastTransitionTime":"2026-01-03T03:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.065438 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.065500 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.065518 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.065546 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.065565 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.169353 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.169412 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.169433 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.169459 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.169478 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.273245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.273327 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.273346 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.273372 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.273394 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.376070 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.376125 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.376145 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.376173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.376192 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.480315 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.480373 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.480391 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.480418 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.480437 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.584428 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.584544 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.584566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.584597 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.584617 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.687896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.687947 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.687960 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.687979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.687990 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.810456 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.810517 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.810531 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.810596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.810610 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.883292 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.883292 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.883301 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:07 crc kubenswrapper[4921]: E0103 03:42:07.883508 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.883590 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:07 crc kubenswrapper[4921]: E0103 03:42:07.883628 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:07 crc kubenswrapper[4921]: E0103 03:42:07.883844 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:07 crc kubenswrapper[4921]: E0103 03:42:07.883992 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.913729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.913829 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.913846 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.913870 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:07 crc kubenswrapper[4921]: I0103 03:42:07.913945 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:07Z","lastTransitionTime":"2026-01-03T03:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.016574 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.016636 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.016647 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.016669 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.016682 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.119389 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.119438 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.119447 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.119469 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.119482 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.222980 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.223038 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.223052 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.223076 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.223086 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.329018 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.329072 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.329082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.329109 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.329122 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.432907 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.432966 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.432996 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.433023 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.433037 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.536189 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.536235 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.536245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.536264 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.536288 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.570643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.570706 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.570722 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.570746 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.570762 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.587741 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:08Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.597169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.597405 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.597516 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.597558 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.597595 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.618622 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:08Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.623394 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.623461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.623475 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.623497 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.623510 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.639981 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:08Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.644955 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.645008 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.645020 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.645042 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.645056 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.660018 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:08Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.663895 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.663932 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.663942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.663960 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.663972 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.678292 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:08Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:08 crc kubenswrapper[4921]: E0103 03:42:08.678478 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.680864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.680907 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.680919 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.680937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.680949 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.783795 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.783858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.783872 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.783894 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.783910 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.886365 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.886422 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.886437 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.886465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.886479 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.990130 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.990211 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.990230 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.990258 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:08 crc kubenswrapper[4921]: I0103 03:42:08.990308 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:08Z","lastTransitionTime":"2026-01-03T03:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.095030 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.095107 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.095126 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.095156 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.095174 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.198256 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.198353 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.198367 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.198388 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.198406 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.301860 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.301931 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.301949 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.301979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.301999 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.405590 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.405643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.405659 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.405686 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.405704 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.509045 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.509112 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.509132 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.509159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.509179 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.613410 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.613551 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.613576 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.613611 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.613635 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.717231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.717297 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.717312 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.717332 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.717348 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.821163 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.821231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.821248 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.821287 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.821303 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.883487 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.883558 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.883596 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.883701 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:09 crc kubenswrapper[4921]: E0103 03:42:09.883888 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:09 crc kubenswrapper[4921]: E0103 03:42:09.884039 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:09 crc kubenswrapper[4921]: E0103 03:42:09.884194 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:09 crc kubenswrapper[4921]: E0103 03:42:09.884339 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.924510 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.924557 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.924566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.924584 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:09 crc kubenswrapper[4921]: I0103 03:42:09.924595 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:09Z","lastTransitionTime":"2026-01-03T03:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.027419 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.027493 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.027511 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.027542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.027562 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.130856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.130955 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.130971 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.130998 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.131016 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.233818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.233884 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.233902 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.233935 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.233958 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.329079 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:10 crc kubenswrapper[4921]: E0103 03:42:10.329332 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:42:10 crc kubenswrapper[4921]: E0103 03:42:10.329463 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:42:42.329426095 +0000 UTC m=+97.940852949 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.338042 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.338099 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.338115 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.338141 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.338158 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.441995 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.442051 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.442069 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.442092 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.442103 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.544401 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.544512 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.544537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.544568 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.544589 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.647123 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.647180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.647193 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.647219 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.647234 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.750501 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.750570 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.750592 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.750620 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.750635 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.853138 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.853230 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.853250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.853319 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.853342 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.955984 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.956055 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.956068 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.956088 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:10 crc kubenswrapper[4921]: I0103 03:42:10.956098 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:10Z","lastTransitionTime":"2026-01-03T03:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.059082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.059131 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.059144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.059162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.059173 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.162541 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.162636 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.162660 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.162697 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.162720 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.264866 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.264949 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.264968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.265000 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.265020 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.368077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.368134 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.368144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.368161 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.368171 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.377052 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/0.log" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.377112 4921 generic.go:334] "Generic (PLEG): container finished" podID="116d8b0a-baa0-4087-9a34-8e890ab8f8dc" containerID="ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f" exitCode=1 Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.377176 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerDied","Data":"ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.378049 4921 scope.go:117] "RemoveContainer" containerID="ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.395178 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.407862 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.421489 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.434505 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.445170 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.457476 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.470529 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.470585 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.470597 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.470621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.470636 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.476145 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.488251 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.501434 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.516441 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.528686 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.541095 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.564879 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.573604 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.573656 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.573667 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.573686 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.573696 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.580920 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.595089 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.606845 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.624372 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:11Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.676968 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.677015 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.677026 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.677046 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.677063 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.780085 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.780130 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.780169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.780205 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.780219 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883563 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883727 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883906 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883946 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883958 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.883986 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:11 crc kubenswrapper[4921]: E0103 03:42:11.884034 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.884060 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.884172 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:11 crc kubenswrapper[4921]: E0103 03:42:11.884389 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:11 crc kubenswrapper[4921]: E0103 03:42:11.884531 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:11 crc kubenswrapper[4921]: E0103 03:42:11.884641 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.987087 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.987167 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.987185 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.987216 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:11 crc kubenswrapper[4921]: I0103 03:42:11.987235 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:11Z","lastTransitionTime":"2026-01-03T03:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.090297 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.090348 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.090362 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.090383 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.090395 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.193050 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.193092 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.193104 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.193124 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.193136 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.295117 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.295196 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.295211 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.295233 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.295248 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.381734 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/0.log" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.381821 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerStarted","Data":"5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.397170 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.397215 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.397227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.397243 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.397256 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.400258 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.416653 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.433871 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.456615 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.472425 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.494696 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.499345 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.499396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.499408 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.499424 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.499436 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.516638 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.535409 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.548813 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.560250 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.575890 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.591910 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.602123 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.602180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.602198 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.602221 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.602241 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.610149 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.622972 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.639643 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.651641 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.660975 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:12Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.705783 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.705824 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.705838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.705857 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.705870 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.808972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.809233 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.809254 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.809322 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.809343 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.911944 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.912017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.912031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.912053 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:12 crc kubenswrapper[4921]: I0103 03:42:12.912068 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:12Z","lastTransitionTime":"2026-01-03T03:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.014559 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.014622 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.014635 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.014655 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.014675 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.118343 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.118424 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.118444 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.118473 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.118493 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.220958 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.221031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.221055 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.221089 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.221112 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.323489 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.323537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.323552 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.323572 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.323583 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.426642 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.426708 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.426731 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.426761 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.426781 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.529551 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.529638 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.529663 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.529694 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.529719 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.632476 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.632561 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.632599 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.632622 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.632639 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.735762 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.735832 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.735847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.735869 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.735886 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.839379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.839453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.839465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.839488 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.839502 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.882847 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.882941 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.882985 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:13 crc kubenswrapper[4921]: E0103 03:42:13.883032 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.883087 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:13 crc kubenswrapper[4921]: E0103 03:42:13.883335 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:13 crc kubenswrapper[4921]: E0103 03:42:13.883387 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:13 crc kubenswrapper[4921]: E0103 03:42:13.883454 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.942685 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.942732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.942747 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.942770 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:13 crc kubenswrapper[4921]: I0103 03:42:13.942783 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:13Z","lastTransitionTime":"2026-01-03T03:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.047352 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.047415 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.047432 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.047461 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.047475 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.149906 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.149964 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.149974 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.149994 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.150010 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.253717 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.253794 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.253818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.253851 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.253878 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.356852 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.356919 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.356937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.356964 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.356984 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.459827 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.459897 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.459915 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.459942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.459966 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.562334 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.562377 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.562387 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.562403 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.562416 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.666360 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.666409 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.666420 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.666438 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.666451 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.769637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.769678 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.769687 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.769703 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.769713 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.872402 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.872455 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.872465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.872482 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.872492 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.912174 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:14Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.929326 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:14Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.945212 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:14Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.967148 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:14Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.983600 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.983992 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.984150 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.984320 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.984609 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:14Z","lastTransitionTime":"2026-01-03T03:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:14 crc kubenswrapper[4921]: I0103 03:42:14.986669 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:14Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.009468 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.029305 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.049163 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.064813 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.078766 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.087724 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.087802 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.087822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.087850 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.087867 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.092872 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.110061 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.124067 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.140938 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.154005 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.165770 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.179417 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:15Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.190589 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.190645 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.190663 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.190707 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.190725 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.293326 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.293755 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.293909 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.294072 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.294207 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.397747 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.397843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.397864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.397925 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.397947 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.501669 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.501718 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.501729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.501746 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.501759 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.605793 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.605849 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.605860 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.605882 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.605897 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.709046 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.709105 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.709123 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.709148 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.709164 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.812423 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.812469 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.812479 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.812496 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.812510 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.883162 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.883227 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.883313 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:15 crc kubenswrapper[4921]: E0103 03:42:15.883386 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.883713 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:15 crc kubenswrapper[4921]: E0103 03:42:15.883761 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:15 crc kubenswrapper[4921]: E0103 03:42:15.883946 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:15 crc kubenswrapper[4921]: E0103 03:42:15.884065 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.884155 4921 scope.go:117] "RemoveContainer" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.915462 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.916204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.916329 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.916454 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:15 crc kubenswrapper[4921]: I0103 03:42:15.916521 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:15Z","lastTransitionTime":"2026-01-03T03:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.018984 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.019340 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.019626 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.019876 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.020117 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.123180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.123223 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.123234 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.123251 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.123262 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.225595 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.225643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.225653 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.225672 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.225687 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.327942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.327972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.327980 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.327994 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.328003 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.395598 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/2.log" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.398336 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.399288 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.417196 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.430729 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.430767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.430777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.430793 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.430803 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.433370 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.447902 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.463640 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.474898 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.507366 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533345 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533698 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533769 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533852 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533929 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.533380 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.547252 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.559566 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.575996 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.591724 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.606293 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.620858 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.632408 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.636043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.636162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.636226 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.636345 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.636411 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.644516 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.656090 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.675093 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:16Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.740090 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.740456 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.740534 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.740603 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.740664 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.843896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.843945 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.843954 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.843975 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.843985 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.946409 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.946456 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.946467 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.946485 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:16 crc kubenswrapper[4921]: I0103 03:42:16.946498 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:16Z","lastTransitionTime":"2026-01-03T03:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.049544 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.049614 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.049629 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.049659 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.049674 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.152774 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.152839 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.152855 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.152881 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.152899 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.256455 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.256522 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.256540 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.256568 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.256593 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.359057 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.359126 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.359139 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.359163 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.359178 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.405643 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/3.log" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.406608 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/2.log" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.410544 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" exitCode=1 Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.410615 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.410685 4921 scope.go:117] "RemoveContainer" containerID="0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.411412 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:42:17 crc kubenswrapper[4921]: E0103 03:42:17.411605 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.440614 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.459677 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.461803 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.461829 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.461838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.461856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.461867 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.478438 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.493127 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.510483 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.525468 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.543237 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.559546 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.566795 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.566865 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.566892 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.566926 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.566954 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.606414 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.632038 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.654646 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.669637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.669692 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.669704 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.669722 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.669736 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.672628 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.688900 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.700865 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.712090 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.729568 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.747822 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0cbecbea896fd983e6cdc4a1cc674a50628c934e3a4506592ff284464b0e6b37\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:41:48Z\\\",\\\"message\\\":\\\"3:41:48.334714 6553 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0103 03:41:48.334714 6553 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0103 03:41:48.334745 6553 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0103 03:41:48.334741 6553 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0103 03:41:48.334770 6553 factory.go:1336] Added *v1.Node event handler 7\\\\nI0103 03:41:48.334798 6553 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0103 03:41:48.334814 6553 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0103 03:41:48.334824 6553 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0103 03:41:48.334871 6553 handler.go:208] Removed *v1.Node event handler 2\\\\nI0103 03:41:48.334873 6553 factory.go:656] Stopping watch factory\\\\nI0103 03:41:48.334892 6553 handler.go:208] Removed *v1.Node event handler 7\\\\nI0103 03:41:48.334913 6553 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0103 03:41:48.335292 6553 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0103 03:41:48.335397 6553 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0103 03:41:48.335463 6553 ovnkube.go:599] Stopped ovnkube\\\\nI0103 03:41:48.335493 6553 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0103 03:41:48.335618 6553 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:17Z\\\",\\\"message\\\":\\\":,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: network-check-target,},ClusterIP:10.217.5.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0103 03:42:16.808392 6933 lb_config.go:1031] Cluster endpoints for openshift-network-diagnostics/network-check-target for network=default are: map[]\\\\nF0103 03:42:16.806060 6933 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal err\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:17Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.772610 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.772648 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.772662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.772681 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.772693 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.875033 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.875072 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.875082 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.875097 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.875130 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.883647 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.883706 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:17 crc kubenswrapper[4921]: E0103 03:42:17.883769 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.883707 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:17 crc kubenswrapper[4921]: E0103 03:42:17.883847 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.883902 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:17 crc kubenswrapper[4921]: E0103 03:42:17.884015 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:17 crc kubenswrapper[4921]: E0103 03:42:17.884081 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.978311 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.978376 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.978396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.978425 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:17 crc kubenswrapper[4921]: I0103 03:42:17.978451 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:17Z","lastTransitionTime":"2026-01-03T03:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.082238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.082338 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.082357 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.082388 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.082408 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.185086 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.185150 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.185171 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.185194 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.185211 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.288179 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.288242 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.288257 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.288301 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.288320 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.391718 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.391803 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.391822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.391854 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.391877 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.415693 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/3.log" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.422053 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.422301 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.446026 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.461005 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.475383 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.493915 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.495442 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.495512 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.495533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.495564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.495585 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.518741 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:17Z\\\",\\\"message\\\":\\\":,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: network-check-target,},ClusterIP:10.217.5.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0103 03:42:16.808392 6933 lb_config.go:1031] Cluster endpoints for openshift-network-diagnostics/network-check-target for network=default are: map[]\\\\nF0103 03:42:16.806060 6933 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal err\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.537422 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.551713 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.571496 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.588374 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.600077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.600126 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.600145 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.600171 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.600190 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.606764 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.619962 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.639182 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.657531 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.677823 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.696894 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.704240 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.704352 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.704370 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.704398 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.704418 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.712246 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.737321 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.808474 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.808542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.808564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.808624 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.808643 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.810742 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.810889 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.810995 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.811204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.811340 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.835381 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.841095 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.841173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.841199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.841231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.841254 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.861887 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.866811 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.866880 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.866905 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.866939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.866964 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.887541 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.892197 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.892258 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.892318 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.892343 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.892362 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.912077 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.917159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.917210 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.917227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.917252 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.917297 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.938252 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:18Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:18 crc kubenswrapper[4921]: E0103 03:42:18.938512 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.941123 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.941179 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.941196 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.941219 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:18 crc kubenswrapper[4921]: I0103 03:42:18.941237 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:18Z","lastTransitionTime":"2026-01-03T03:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.044454 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.044571 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.044588 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.044607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.044619 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.148128 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.148183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.148199 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.148222 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.148239 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.252098 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.252182 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.252202 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.252233 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.252254 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.354939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.355002 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.355017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.355041 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.355055 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.457646 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.457689 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.457699 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.457714 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.457724 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.559892 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.559952 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.559961 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.559975 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.559986 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.663117 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.663189 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.663208 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.663238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.663262 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.767169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.767307 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.767338 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.767374 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.767397 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.870996 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.871076 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.871096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.871128 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.871152 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.883155 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.883206 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.883235 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.883315 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:19 crc kubenswrapper[4921]: E0103 03:42:19.883535 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:19 crc kubenswrapper[4921]: E0103 03:42:19.883709 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:19 crc kubenswrapper[4921]: E0103 03:42:19.883881 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:19 crc kubenswrapper[4921]: E0103 03:42:19.883975 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.973970 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.974056 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.974077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.974106 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:19 crc kubenswrapper[4921]: I0103 03:42:19.974127 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:19Z","lastTransitionTime":"2026-01-03T03:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.078493 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.078598 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.078626 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.078659 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.078687 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.182395 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.182512 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.182533 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.182560 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.182580 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.286493 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.286558 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.286575 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.286610 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.286633 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.390144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.390213 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.390232 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.390261 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.390309 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.493891 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.493945 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.493965 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.493989 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.494009 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.597158 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.597456 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.597736 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.597796 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.597818 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.701532 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.701605 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.701624 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.701701 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.701768 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.806042 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.806119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.806134 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.806154 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.806167 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.909815 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.909899 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.909926 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.909963 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:20 crc kubenswrapper[4921]: I0103 03:42:20.909982 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:20Z","lastTransitionTime":"2026-01-03T03:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.013566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.013623 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.013639 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.013662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.013678 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.117488 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.117542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.117554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.117573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.117591 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.221112 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.221164 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.221180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.221200 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.221210 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.325074 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.325143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.325162 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.325189 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.325208 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.428474 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.428536 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.428553 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.428582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.428600 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.532214 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.532317 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.532337 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.532366 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.532387 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.644320 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.644483 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.644506 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.644570 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.644593 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.748561 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.748618 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.748627 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.748647 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.748659 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.862923 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.863096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.863117 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.863180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.863234 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.883305 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.883338 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.883316 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.883434 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:21 crc kubenswrapper[4921]: E0103 03:42:21.883567 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:21 crc kubenswrapper[4921]: E0103 03:42:21.883706 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:21 crc kubenswrapper[4921]: E0103 03:42:21.883868 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:21 crc kubenswrapper[4921]: E0103 03:42:21.883988 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.967735 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.967813 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.967831 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.967859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:21 crc kubenswrapper[4921]: I0103 03:42:21.967877 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:21Z","lastTransitionTime":"2026-01-03T03:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.071484 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.071566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.071584 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.071618 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.071638 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.175482 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.175544 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.175560 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.175583 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.175600 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.279389 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.279460 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.279477 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.279498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.279510 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.383547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.383632 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.383658 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.383688 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.383712 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.487777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.487853 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.487873 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.487904 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.487928 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.591811 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.591885 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.591904 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.591935 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.591957 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.696207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.696265 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.696298 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.696316 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.696330 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.800427 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.800510 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.800528 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.800561 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.800582 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.904014 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.904097 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.904116 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.904141 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:22 crc kubenswrapper[4921]: I0103 03:42:22.904159 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:22Z","lastTransitionTime":"2026-01-03T03:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.007563 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.007619 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.007636 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.007662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.007680 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.111022 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.111213 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.111238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.111266 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.111321 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.214832 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.214934 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.214961 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.214996 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.215016 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.318016 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.318054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.318062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.318081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.318092 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.421634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.421715 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.421731 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.421763 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.421784 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.525642 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.525788 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.525813 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.525845 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.525865 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.629602 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.629684 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.629702 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.629735 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.629756 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.733549 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.733631 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.733661 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.733701 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.733729 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.837990 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.838083 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.838101 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.838133 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.838153 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.883388 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.883427 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.883441 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.883441 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:23 crc kubenswrapper[4921]: E0103 03:42:23.883661 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:23 crc kubenswrapper[4921]: E0103 03:42:23.883835 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:23 crc kubenswrapper[4921]: E0103 03:42:23.884088 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:23 crc kubenswrapper[4921]: E0103 03:42:23.884197 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.941820 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.941898 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.941925 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.941960 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:23 crc kubenswrapper[4921]: I0103 03:42:23.941982 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:23Z","lastTransitionTime":"2026-01-03T03:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.045781 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.045860 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.045878 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.045906 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.045925 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.149253 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.149355 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.149376 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.149412 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.149436 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.252741 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.252822 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.252847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.252884 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.252915 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.357009 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.357070 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.357085 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.357109 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.357127 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.460523 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.460634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.460657 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.460688 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.460708 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.564801 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.564889 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.564909 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.564942 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.564965 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.668339 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.668428 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.668448 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.668479 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.668501 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.772643 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.772732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.772750 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.772777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.772798 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.876672 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.876762 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.876790 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.876832 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.876863 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.901924 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.909796 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-6xvww" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"165b4d4c-dd8c-4c0f-8a6c-8282a934fb4f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edb21c0d0e831d5316c97d632b042d46583b9f8aac4cc64dc96741265c77ef6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b70e822bcdc40aa2be8f4eb48784156a168a40c6c7b7f15fa162b39f8d590b36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54a4a4b7ff4c18bdbf1a775911b916edba725a0d9c9c62676e3a44304cd9343d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5d0af218428cc4a8573fab87fd82976d5bf415516393bcff3df665358b4fca24\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://789f1c1963af2baf2d87124db9f820e432ef98c1cf23450a7eefebf9918d2f04\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://76ee54a4cca42e256fa001124acea936ebca3019449faa532bf2e635b1f8554e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc903d4efd9a091eb783766d27c72cb85c5f7a7f8095a75734549d0e810d1f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9t6xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-6xvww\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:24Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.946210 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf34943c-bfe4-4411-af8a-189b14e35a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:17Z\\\",\\\"message\\\":\\\":,Protocol:TCP,Port:80,TargetPort:{0 8080 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{app: network-check-target,},ClusterIP:10.217.5.219,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.219],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI0103 03:42:16.808392 6933 lb_config.go:1031] Cluster endpoints for openshift-network-diagnostics/network-check-target for network=default are: map[]\\\\nF0103 03:42:16.806060 6933 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal err\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mgv5c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbb7k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:24Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.973322 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f16567fe-14ec-443c-966a-78e2e77f48fe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\":/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1767411667\\\\\\\\\\\\\\\" (2026-01-03 03:41:06 +0000 UTC to 2026-02-02 03:41:07 +0000 UTC (now=2026-01-03 03:41:23.393761005 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394023 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1767411678\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1767411678\\\\\\\\\\\\\\\" (2026-01-03 02:41:17 +0000 UTC to 2027-01-03 02:41:17 +0000 UTC (now=2026-01-03 03:41:23.394000202 +0000 UTC))\\\\\\\"\\\\nI0103 03:41:23.394075 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0103 03:41:23.394097 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0103 03:41:23.394116 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394141 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0103 03:41:23.394176 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-804878105/tls.crt::/tmp/serving-cert-804878105/tls.key\\\\\\\"\\\\nI0103 03:41:23.394345 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF0103 03:41:23.395713 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0103 03:41:23.397316 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0103 03:41:23.397347 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0103 03:41:23.399167 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0103 03:41:23.399194 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:24Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.978908 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.978950 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.978961 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.978982 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.978998 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:24Z","lastTransitionTime":"2026-01-03T03:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:24 crc kubenswrapper[4921]: I0103 03:42:24.997291 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://450cbd8e1a520eb873a9c117a3064adbd80c66f89a34eb878c89c85ce8e8e070\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:24Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.016375 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mhxl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c62912da-e574-4877-9c1a-59e2a10c1d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aacc20450ff036b401843506563b5b8eb175adc9db39763a2ec04eb2510fe742\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t4ck4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mhxl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.034822 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-vsp88" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snpwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-vsp88\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.056761 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"563c742d-72f3-4363-9dab-b61cf1ab3141\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85f7c30ed0957fe1749d1cdc171f2c70a7cc2960703c2170b37412f39d796aee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8605f82d726cddf6dfe1990024a1a02c06cb416169d6ecfd8b92525e8e0624e6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b4c2480190bc4abc93346a34c376a3149c3a3870dfc4d3938632b02441420dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.074253 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.082207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.082265 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.082295 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.082322 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.082342 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.099962 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f22b9700c705e3f558b9bf0c7405a136478f38d4c5af8a7e02d1f7d11df3f545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dff2a8f7744a7f6983d20d5f1d80e9c48cdb716898036bebcd6c8269cead0e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.123010 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.141765 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faa96c8e-1d8c-4ff1-80e9-f6fed0bb7bfc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a38140c40f219944b3ad4a9ee510914ed9cb9ceb99ba4c850794c6233ea6de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6407956ae3b37c2427dfa9013be072de07eb44dbd928c3250662a415faf061ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76d722f09a43568e43669e987504ee96bc98dfe82d6357071770d2eaee54e9e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48bd5f78a7cd268b4a7151c143ac7cb125e606bb757b55997050c358fd43dc5e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-03T03:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:04Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.157718 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-qnm8g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eed39a0f-cb1e-451a-a36c-233b20b3ca1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c67ecb847e1a112d6a28812bd893db9cdee080283cdb8c04fa12aeccb624a696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5qtxx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:23Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-qnm8g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.170675 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"429ab47e-68f8-4b60-aa4c-ab79a764b7db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4ebd30cfeb82e5e587ea074e84efd0f4fa9cb6479c4b11d4644129e42de46d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xm5sd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-cctxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.185797 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.185829 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.185838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.185853 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.185863 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.190475 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-666ct" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"116d8b0a-baa0-4087-9a34-8e890ab8f8dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-03T03:42:11Z\\\",\\\"message\\\":\\\"2026-01-03T03:41:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f\\\\n2026-01-03T03:41:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d6678006-7f15-4200-a85e-45f607fa475f to /host/opt/cni/bin/\\\\n2026-01-03T03:41:26Z [verbose] multus-daemon started\\\\n2026-01-03T03:41:26Z [verbose] Readiness Indicator file check\\\\n2026-01-03T03:42:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-03T03:41:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mrslm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-666ct\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.210870 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bf98706c-cd2f-4ab8-98aa-ed60ce7dd404\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e106c53befa5dfcf17ddda3a32f557c9517965920d324dc6ce16815761485b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33826fed02b8e59834fdca4a152837da743e0433f85d3cbf78717f019dd50511\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q8cq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-03T03:41:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cnc2p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.233239 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65702a69d212de03482eee5984ed6b565a41c48b115c30e586d85186cf4aee45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-03T03:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.254906 4921 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-03T03:41:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:25Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.289228 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.289325 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.289342 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.289371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.289389 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.392795 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.392844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.392856 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.392876 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.392891 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.496149 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.496208 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.496221 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.496241 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.496256 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.600524 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.600621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.600638 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.600663 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.600683 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.703532 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.703586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.703596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.703613 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.703624 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.807183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.807238 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.807253 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.807310 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.807328 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.883164 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.883212 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.883170 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.883334 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:25 crc kubenswrapper[4921]: E0103 03:42:25.883367 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:25 crc kubenswrapper[4921]: E0103 03:42:25.883565 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:25 crc kubenswrapper[4921]: E0103 03:42:25.883666 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:25 crc kubenswrapper[4921]: E0103 03:42:25.883779 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.910017 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.910067 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.910077 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.910094 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:25 crc kubenswrapper[4921]: I0103 03:42:25.910105 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:25Z","lastTransitionTime":"2026-01-03T03:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.012644 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.012705 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.012725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.012752 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.012771 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.115997 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.116066 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.116084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.116109 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.116131 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.219762 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.219833 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.219850 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.219881 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.219901 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.323308 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.323379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.323397 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.323426 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.323447 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.427232 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.427330 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.427352 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.427378 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.427397 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.530789 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.530837 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.530855 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.530876 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.530889 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.634542 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.634631 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.634653 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.634679 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.634699 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.738350 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.738403 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.738413 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.738434 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.738447 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.842022 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.842093 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.842113 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.842141 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.842162 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.945155 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.945204 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.945219 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.945240 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:26 crc kubenswrapper[4921]: I0103 03:42:26.945255 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:26Z","lastTransitionTime":"2026-01-03T03:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.050160 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.050246 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.050264 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.050352 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.050371 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.152583 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.152645 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.152655 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.152670 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.152682 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.255093 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.255159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.255181 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.255208 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.255230 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.358656 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.358721 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.358747 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.358779 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.358801 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.461761 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.461806 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.461818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.461835 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.461850 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.565478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.565547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.565567 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.565595 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.565616 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.669300 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.669359 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.669379 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.669406 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.669425 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.730686 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.730798 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.730898 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.730981 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.730947881 +0000 UTC m=+147.342374745 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.730999 4921 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.731020 4921 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.731079 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.731064574 +0000 UTC m=+147.342491408 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.731102 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.731091585 +0000 UTC m=+147.342518419 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.773183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.773250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.773338 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.773368 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.773388 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.832515 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.832611 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.832833 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.832864 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.832881 4921 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.832971 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.832946738 +0000 UTC m=+147.444373572 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.833037 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.833061 4921 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.833078 4921 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.833160 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.833110523 +0000 UTC m=+147.444537357 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.877044 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.877096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.877112 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.877135 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.877150 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.883679 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.883707 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.883691 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.883800 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.883798 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.883955 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.884086 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:27 crc kubenswrapper[4921]: E0103 03:42:27.884140 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.980486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.980560 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.980586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.980621 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:27 crc kubenswrapper[4921]: I0103 03:42:27.980644 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:27Z","lastTransitionTime":"2026-01-03T03:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.084843 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.085435 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.085472 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.085508 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.085548 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.187721 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.187770 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.187780 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.187797 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.187811 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.291063 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.291146 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.291172 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.291207 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.291232 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.396548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.396635 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.396662 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.396696 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.396721 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.500558 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.500635 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.500654 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.500702 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.500722 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.604864 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.604940 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.604960 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.604991 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.605013 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.708805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.708858 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.708873 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.708893 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.708910 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.812661 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.812721 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.812736 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.812757 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.812771 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.916378 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.916450 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.916477 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.916514 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:28 crc kubenswrapper[4921]: I0103 03:42:28.916536 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:28Z","lastTransitionTime":"2026-01-03T03:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.019939 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.020029 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.020043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.020062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.020075 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.093020 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.093113 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.093139 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.093169 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.093191 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.115046 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.122915 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.123073 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.123103 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.123167 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.123192 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.146222 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.153675 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.153730 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.153748 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.153775 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.153796 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.173873 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.180170 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.180313 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.180335 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.180360 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.180382 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.202928 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.211299 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.211369 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.211389 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.211416 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.211438 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.234534 4921 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-03T03:42:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"904f1a66-3c4f-4ba4-86d9-b5f3204a6dea\\\",\\\"systemUUID\\\":\\\"3994d32f-6f26-48b7-a835-33b55e41bd4d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-03T03:42:29Z is after 2025-08-24T17:21:41Z" Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.234977 4921 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.238572 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.238646 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.238671 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.238706 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.238729 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.341962 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.342013 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.342030 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.342054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.342069 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.445906 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.445982 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.446000 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.446028 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.446048 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.549509 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.549568 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.549586 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.549607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.549622 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.653108 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.653191 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.653215 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.653250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.653326 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.756784 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.756859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.756877 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.756903 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.756922 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.861973 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.862086 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.862110 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.862148 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.862185 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.882967 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.883072 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.883106 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.883085 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.883539 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.883706 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.883859 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:29 crc kubenswrapper[4921]: E0103 03:42:29.884037 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.965952 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.966035 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.966053 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.966084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:29 crc kubenswrapper[4921]: I0103 03:42:29.966102 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:29Z","lastTransitionTime":"2026-01-03T03:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.069975 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.070062 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.070100 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.070138 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.070164 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.173963 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.174050 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.174068 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.174097 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.174116 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.277828 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.277905 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.277923 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.277952 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.277975 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.382066 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.382138 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.382166 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.382203 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.382229 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.484764 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.484827 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.484845 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.484870 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.484887 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.589147 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.589360 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.589396 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.589429 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.589449 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.692543 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.692610 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.692634 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.692667 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.692689 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.795480 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.795547 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.795564 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.795592 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.795611 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.885597 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:42:30 crc kubenswrapper[4921]: E0103 03:42:30.885970 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.898596 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.898681 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.898699 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.898732 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:30 crc kubenswrapper[4921]: I0103 03:42:30.898752 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:30Z","lastTransitionTime":"2026-01-03T03:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.002633 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.002756 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.002777 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.002845 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.002871 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.106540 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.106600 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.106615 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.106639 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.106653 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.210193 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.210255 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.210307 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.210333 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.210352 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.313013 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.313085 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.313097 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.313121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.313137 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.416486 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.416561 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.416582 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.416607 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.416625 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.519453 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.519509 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.519520 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.519540 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.519557 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.622755 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.622804 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.622814 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.622837 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.622851 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.726885 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.726932 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.726944 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.726961 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.726974 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.832029 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.832103 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.832121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.832148 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.832173 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.882703 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.882776 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.882817 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:31 crc kubenswrapper[4921]: E0103 03:42:31.882952 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.882976 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:31 crc kubenswrapper[4921]: E0103 03:42:31.883127 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:31 crc kubenswrapper[4921]: E0103 03:42:31.883235 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:31 crc kubenswrapper[4921]: E0103 03:42:31.883319 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.936592 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.936674 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.936693 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.936726 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:31 crc kubenswrapper[4921]: I0103 03:42:31.936745 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:31Z","lastTransitionTime":"2026-01-03T03:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.040009 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.040084 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.040101 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.040130 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.040151 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.144362 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.144450 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.144473 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.144513 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.144544 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.248025 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.248133 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.248157 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.248188 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.248207 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.350479 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.350553 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.350573 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.350597 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.350617 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.454121 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.454167 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.454177 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.454196 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.454210 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.557335 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.557382 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.557393 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.557410 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.557423 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.661602 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.661689 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.661709 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.661737 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.661757 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.765689 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.765765 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.765789 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.765818 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.765838 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.869788 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.869881 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.869901 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.869932 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.869952 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.972243 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.972330 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.972349 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.972371 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:32 crc kubenswrapper[4921]: I0103 03:42:32.972391 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:32Z","lastTransitionTime":"2026-01-03T03:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.075737 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.075815 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.075833 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.075862 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.075879 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.178713 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.178767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.178780 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.178802 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.178825 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.282483 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.282566 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.282584 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.282615 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.282637 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.386412 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.386498 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.386520 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.386554 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.386575 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.489905 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.489979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.490005 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.490038 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.490062 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.593847 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.593929 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.593952 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.593985 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.594009 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.697680 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.697749 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.697766 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.697794 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.697811 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.801791 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.801999 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.802025 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.802436 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.802472 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.883048 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.883149 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.883215 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.883089 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:33 crc kubenswrapper[4921]: E0103 03:42:33.883340 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:33 crc kubenswrapper[4921]: E0103 03:42:33.883438 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:33 crc kubenswrapper[4921]: E0103 03:42:33.883553 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:33 crc kubenswrapper[4921]: E0103 03:42:33.883678 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.906502 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.906570 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.906587 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.906617 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:33 crc kubenswrapper[4921]: I0103 03:42:33.906638 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:33Z","lastTransitionTime":"2026-01-03T03:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.009767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.009844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.009865 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.009896 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.009914 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.114088 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.114165 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.114183 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.114209 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.114228 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.218364 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.218541 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.218569 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.218667 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.218820 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.323767 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.323826 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.323838 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.323861 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.323873 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.427271 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.427356 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.427374 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.427400 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.427420 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.530511 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.530631 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.530694 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.530725 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.530743 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.632813 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.632890 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.632914 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.632949 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.632974 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.735788 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.735862 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.735885 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.735912 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.735933 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.839537 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.839625 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.839646 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.839671 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.839689 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.906150 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.906126106 podStartE2EDuration="1m9.906126106s" podCreationTimestamp="2026-01-03 03:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:34.906054374 +0000 UTC m=+90.517481238" watchObservedRunningTime="2026-01-03 03:42:34.906126106 +0000 UTC m=+90.517552960" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.943019 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.943361 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.943434 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.943535 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:34 crc kubenswrapper[4921]: I0103 03:42:34.943559 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:34Z","lastTransitionTime":"2026-01-03T03:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.029926 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=37.029900892 podStartE2EDuration="37.029900892s" podCreationTimestamp="2026-01-03 03:41:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.0034624 +0000 UTC m=+90.614889264" watchObservedRunningTime="2026-01-03 03:42:35.029900892 +0000 UTC m=+90.641327726" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.047209 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.047683 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.047703 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.047721 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.047765 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.055304 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qnm8g" podStartSLOduration=72.055221624 podStartE2EDuration="1m12.055221624s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.030494517 +0000 UTC m=+90.641921351" watchObservedRunningTime="2026-01-03 03:42:35.055221624 +0000 UTC m=+90.666648458" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.088854 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cnc2p" podStartSLOduration=71.088831136 podStartE2EDuration="1m11.088831136s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.072980676 +0000 UTC m=+90.684407510" watchObservedRunningTime="2026-01-03 03:42:35.088831136 +0000 UTC m=+90.700257960" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.089251 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=11.089245317 podStartE2EDuration="11.089245317s" podCreationTimestamp="2026-01-03 03:42:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.088473327 +0000 UTC m=+90.699900151" watchObservedRunningTime="2026-01-03 03:42:35.089245317 +0000 UTC m=+90.700672141" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.150354 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.150409 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.150422 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.150440 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.150453 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.173344 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podStartSLOduration=72.173315309 podStartE2EDuration="1m12.173315309s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.145964274 +0000 UTC m=+90.757391108" watchObservedRunningTime="2026-01-03 03:42:35.173315309 +0000 UTC m=+90.784742143" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.194897 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-666ct" podStartSLOduration=72.194874252 podStartE2EDuration="1m12.194874252s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.174143772 +0000 UTC m=+90.785570606" watchObservedRunningTime="2026-01-03 03:42:35.194874252 +0000 UTC m=+90.806301086" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.214706 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.214679237 podStartE2EDuration="1m11.214679237s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.195750105 +0000 UTC m=+90.807176949" watchObservedRunningTime="2026-01-03 03:42:35.214679237 +0000 UTC m=+90.826106081" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.231306 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-mhxl4" podStartSLOduration=72.231261198 podStartE2EDuration="1m12.231261198s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.230855607 +0000 UTC m=+90.842282441" watchObservedRunningTime="2026-01-03 03:42:35.231261198 +0000 UTC m=+90.842688032" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.252940 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.252978 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.252990 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.253009 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.253022 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.288011 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-6xvww" podStartSLOduration=72.287985943 podStartE2EDuration="1m12.287985943s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:35.255855181 +0000 UTC m=+90.867282015" watchObservedRunningTime="2026-01-03 03:42:35.287985943 +0000 UTC m=+90.899412777" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.356182 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.356258 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.356310 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.356335 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.356353 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.459556 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.459889 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.459999 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.460115 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.460204 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.563865 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.563928 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.563945 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.563969 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.563986 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.667711 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.667770 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.667787 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.667816 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.667834 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.772031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.772101 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.772119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.772148 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.772169 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.876140 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.876211 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.876231 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.876258 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.876314 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.883561 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.883563 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.883623 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.883734 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:35 crc kubenswrapper[4921]: E0103 03:42:35.883905 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:35 crc kubenswrapper[4921]: E0103 03:42:35.884061 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:35 crc kubenswrapper[4921]: E0103 03:42:35.884415 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:35 crc kubenswrapper[4921]: E0103 03:42:35.884629 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.980173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.980235 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.980247 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.980291 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:35 crc kubenswrapper[4921]: I0103 03:42:35.980305 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:35Z","lastTransitionTime":"2026-01-03T03:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.087158 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.087245 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.087316 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.087376 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.087412 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.191882 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.191972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.191996 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.192029 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.192053 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.296054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.296548 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.296704 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.296862 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.297077 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.401684 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.401749 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.401773 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.401805 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.401828 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.506144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.506696 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.506930 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.507118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.507365 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.610489 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.610565 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.610587 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.610616 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.610641 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.714234 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.714340 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.714366 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.714407 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.714434 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.820040 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.820105 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.820119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.820142 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.820236 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.924557 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.924628 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.924646 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.924675 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:36 crc kubenswrapper[4921]: I0103 03:42:36.924699 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:36Z","lastTransitionTime":"2026-01-03T03:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.028742 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.028830 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.028859 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.028892 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.028915 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.133111 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.133180 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.133200 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.133227 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.133246 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.236415 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.236481 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.236501 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.236529 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.236548 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.340866 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.340937 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.340960 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.340997 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.341021 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.444159 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.444250 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.444306 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.444344 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.444368 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.547887 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.547972 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.547997 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.548031 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.548055 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.651501 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.651583 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.651605 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.651638 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.651659 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.755727 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.755819 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.755844 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.755885 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.755917 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.859096 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.859175 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.859200 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.859236 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.859259 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.883221 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.883305 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.883371 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.883219 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:37 crc kubenswrapper[4921]: E0103 03:42:37.883503 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:37 crc kubenswrapper[4921]: E0103 03:42:37.883676 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:37 crc kubenswrapper[4921]: E0103 03:42:37.883799 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:37 crc kubenswrapper[4921]: E0103 03:42:37.883935 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.963443 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.963506 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.963520 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.963541 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:37 crc kubenswrapper[4921]: I0103 03:42:37.963555 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:37Z","lastTransitionTime":"2026-01-03T03:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.067054 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.067123 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.067144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.067171 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.067188 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.170398 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.170462 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.170482 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.170508 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.170532 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.273494 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.273591 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.273609 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.273637 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.273658 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.376951 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.377040 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.377066 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.377101 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.377131 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.479971 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.480081 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.480106 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.480541 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.480844 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.584347 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.584420 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.584438 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.584465 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.584484 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.687644 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.687699 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.687718 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.687741 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.687759 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.791069 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.791119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.791140 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.791173 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.791197 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.894579 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.894653 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.894671 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.894695 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.894716 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.998118 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.998195 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.998222 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.998254 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:38 crc kubenswrapper[4921]: I0103 03:42:38.998331 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:38Z","lastTransitionTime":"2026-01-03T03:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.102127 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.102184 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.102202 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.102230 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.102249 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.205979 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.206119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.206143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.206174 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.206196 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.308853 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.308919 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.308933 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.308955 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.308969 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.413099 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.413143 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.413156 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.413176 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.413192 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.516422 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.516468 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.516478 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.516494 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.516506 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.579043 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.579105 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.579119 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.579144 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.579162 4921 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-03T03:42:39Z","lastTransitionTime":"2026-01-03T03:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.641954 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt"] Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.642384 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.644308 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.644367 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.645179 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.648157 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.794082 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f248fa-629e-480b-878b-85aec56e9ecf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.794193 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6f248fa-629e-480b-878b-85aec56e9ecf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.794316 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.794422 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.794519 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6f248fa-629e-480b-878b-85aec56e9ecf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.883559 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.883567 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.883645 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.883719 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:39 crc kubenswrapper[4921]: E0103 03:42:39.883932 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:39 crc kubenswrapper[4921]: E0103 03:42:39.884041 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:39 crc kubenswrapper[4921]: E0103 03:42:39.884159 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:39 crc kubenswrapper[4921]: E0103 03:42:39.884364 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.895892 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6f248fa-629e-480b-878b-85aec56e9ecf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.895962 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.895998 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.896023 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6f248fa-629e-480b-878b-85aec56e9ecf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.896056 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f248fa-629e-480b-878b-85aec56e9ecf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.896143 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.896222 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/e6f248fa-629e-480b-878b-85aec56e9ecf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.897889 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e6f248fa-629e-480b-878b-85aec56e9ecf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.909661 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6f248fa-629e-480b-878b-85aec56e9ecf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.916481 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6f248fa-629e-480b-878b-85aec56e9ecf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6fgqt\" (UID: \"e6f248fa-629e-480b-878b-85aec56e9ecf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:39 crc kubenswrapper[4921]: I0103 03:42:39.955242 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" Jan 03 03:42:40 crc kubenswrapper[4921]: I0103 03:42:40.511705 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" event={"ID":"e6f248fa-629e-480b-878b-85aec56e9ecf","Type":"ContainerStarted","Data":"db29aa1a243190c2bd316112e9e21b69ded77fc96d5c5a259727a8cd070676a6"} Jan 03 03:42:40 crc kubenswrapper[4921]: I0103 03:42:40.512024 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" event={"ID":"e6f248fa-629e-480b-878b-85aec56e9ecf","Type":"ContainerStarted","Data":"85bc8ecebf49a0e2e989402583ef74a2834c9867d6c582d73cb904b8e4422a06"} Jan 03 03:42:40 crc kubenswrapper[4921]: I0103 03:42:40.538617 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6fgqt" podStartSLOduration=77.538597838 podStartE2EDuration="1m17.538597838s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:40.537154979 +0000 UTC m=+96.148581813" watchObservedRunningTime="2026-01-03 03:42:40.538597838 +0000 UTC m=+96.150024682" Jan 03 03:42:41 crc kubenswrapper[4921]: I0103 03:42:41.883803 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:41 crc kubenswrapper[4921]: I0103 03:42:41.883859 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:41 crc kubenswrapper[4921]: I0103 03:42:41.883886 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:41 crc kubenswrapper[4921]: E0103 03:42:41.884025 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:41 crc kubenswrapper[4921]: I0103 03:42:41.884054 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:41 crc kubenswrapper[4921]: E0103 03:42:41.884159 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:41 crc kubenswrapper[4921]: E0103 03:42:41.884260 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:41 crc kubenswrapper[4921]: E0103 03:42:41.884374 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:42 crc kubenswrapper[4921]: I0103 03:42:42.425941 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:42 crc kubenswrapper[4921]: E0103 03:42:42.426333 4921 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:42:42 crc kubenswrapper[4921]: E0103 03:42:42.426439 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs podName:fb5f46fd-64d0-4308-8fc6-5eff70d2521b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:46.426406695 +0000 UTC m=+162.037833559 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs") pod "network-metrics-daemon-vsp88" (UID: "fb5f46fd-64d0-4308-8fc6-5eff70d2521b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 03 03:42:42 crc kubenswrapper[4921]: I0103 03:42:42.884413 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:42:42 crc kubenswrapper[4921]: E0103 03:42:42.885710 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbb7k_openshift-ovn-kubernetes(bf34943c-bfe4-4411-af8a-189b14e35a82)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" Jan 03 03:42:43 crc kubenswrapper[4921]: I0103 03:42:43.882864 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:43 crc kubenswrapper[4921]: I0103 03:42:43.883503 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:43 crc kubenswrapper[4921]: I0103 03:42:43.883564 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:43 crc kubenswrapper[4921]: I0103 03:42:43.883630 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:43 crc kubenswrapper[4921]: E0103 03:42:43.883741 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:43 crc kubenswrapper[4921]: E0103 03:42:43.883969 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:43 crc kubenswrapper[4921]: E0103 03:42:43.884072 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:43 crc kubenswrapper[4921]: E0103 03:42:43.884157 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:45 crc kubenswrapper[4921]: I0103 03:42:45.883052 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:45 crc kubenswrapper[4921]: E0103 03:42:45.883806 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:45 crc kubenswrapper[4921]: I0103 03:42:45.883139 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:45 crc kubenswrapper[4921]: E0103 03:42:45.884003 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:45 crc kubenswrapper[4921]: I0103 03:42:45.883252 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:45 crc kubenswrapper[4921]: E0103 03:42:45.884183 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:45 crc kubenswrapper[4921]: I0103 03:42:45.883079 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:45 crc kubenswrapper[4921]: E0103 03:42:45.884388 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:45 crc kubenswrapper[4921]: I0103 03:42:45.907849 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 03 03:42:47 crc kubenswrapper[4921]: I0103 03:42:47.883017 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:47 crc kubenswrapper[4921]: I0103 03:42:47.883125 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:47 crc kubenswrapper[4921]: I0103 03:42:47.883171 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:47 crc kubenswrapper[4921]: I0103 03:42:47.883125 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:47 crc kubenswrapper[4921]: E0103 03:42:47.883554 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:47 crc kubenswrapper[4921]: E0103 03:42:47.883721 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:47 crc kubenswrapper[4921]: E0103 03:42:47.883798 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:47 crc kubenswrapper[4921]: E0103 03:42:47.883966 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:49 crc kubenswrapper[4921]: I0103 03:42:49.882989 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:49 crc kubenswrapper[4921]: I0103 03:42:49.883021 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:49 crc kubenswrapper[4921]: I0103 03:42:49.883082 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:49 crc kubenswrapper[4921]: E0103 03:42:49.883358 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:49 crc kubenswrapper[4921]: I0103 03:42:49.883485 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:49 crc kubenswrapper[4921]: E0103 03:42:49.883714 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:49 crc kubenswrapper[4921]: E0103 03:42:49.883985 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:49 crc kubenswrapper[4921]: E0103 03:42:49.884130 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:51 crc kubenswrapper[4921]: I0103 03:42:51.883576 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:51 crc kubenswrapper[4921]: I0103 03:42:51.883643 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:51 crc kubenswrapper[4921]: I0103 03:42:51.883781 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:51 crc kubenswrapper[4921]: I0103 03:42:51.883643 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:51 crc kubenswrapper[4921]: E0103 03:42:51.883973 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:51 crc kubenswrapper[4921]: E0103 03:42:51.884138 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:51 crc kubenswrapper[4921]: E0103 03:42:51.884343 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:51 crc kubenswrapper[4921]: E0103 03:42:51.884574 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:53 crc kubenswrapper[4921]: I0103 03:42:53.882795 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:53 crc kubenswrapper[4921]: I0103 03:42:53.882847 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:53 crc kubenswrapper[4921]: I0103 03:42:53.882862 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:53 crc kubenswrapper[4921]: I0103 03:42:53.882790 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:53 crc kubenswrapper[4921]: E0103 03:42:53.883004 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:53 crc kubenswrapper[4921]: E0103 03:42:53.883213 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:53 crc kubenswrapper[4921]: E0103 03:42:53.883344 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:53 crc kubenswrapper[4921]: E0103 03:42:53.883443 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:54 crc kubenswrapper[4921]: I0103 03:42:54.926226 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.926207222 podStartE2EDuration="9.926207222s" podCreationTimestamp="2026-01-03 03:42:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:54.924601308 +0000 UTC m=+110.536028152" watchObservedRunningTime="2026-01-03 03:42:54.926207222 +0000 UTC m=+110.537634046" Jan 03 03:42:55 crc kubenswrapper[4921]: I0103 03:42:55.883387 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:55 crc kubenswrapper[4921]: I0103 03:42:55.883504 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:55 crc kubenswrapper[4921]: I0103 03:42:55.883512 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:55 crc kubenswrapper[4921]: E0103 03:42:55.883700 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:55 crc kubenswrapper[4921]: I0103 03:42:55.883766 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:55 crc kubenswrapper[4921]: E0103 03:42:55.883915 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:55 crc kubenswrapper[4921]: E0103 03:42:55.884056 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:55 crc kubenswrapper[4921]: E0103 03:42:55.884171 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.573913 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/1.log" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.575063 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/0.log" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.575153 4921 generic.go:334] "Generic (PLEG): container finished" podID="116d8b0a-baa0-4087-9a34-8e890ab8f8dc" containerID="5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b" exitCode=1 Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.575200 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerDied","Data":"5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b"} Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.575350 4921 scope.go:117] "RemoveContainer" containerID="ce50199038c0b582ad52d07f1094415aea2b641bcacdc86942125e5aea0d271f" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.575828 4921 scope.go:117] "RemoveContainer" containerID="5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b" Jan 03 03:42:57 crc kubenswrapper[4921]: E0103 03:42:57.576037 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-666ct_openshift-multus(116d8b0a-baa0-4087-9a34-8e890ab8f8dc)\"" pod="openshift-multus/multus-666ct" podUID="116d8b0a-baa0-4087-9a34-8e890ab8f8dc" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.882962 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.883011 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.882962 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.882962 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:57 crc kubenswrapper[4921]: E0103 03:42:57.883125 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:57 crc kubenswrapper[4921]: E0103 03:42:57.883254 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:57 crc kubenswrapper[4921]: E0103 03:42:57.883841 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:42:57 crc kubenswrapper[4921]: E0103 03:42:57.883959 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:57 crc kubenswrapper[4921]: I0103 03:42:57.884028 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.580868 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/1.log" Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.584860 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/3.log" Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.588579 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerStarted","Data":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.589085 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.619318 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podStartSLOduration=94.619259596 podStartE2EDuration="1m34.619259596s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:42:58.619208005 +0000 UTC m=+114.230634879" watchObservedRunningTime="2026-01-03 03:42:58.619259596 +0000 UTC m=+114.230686430" Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.715812 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vsp88"] Jan 03 03:42:58 crc kubenswrapper[4921]: I0103 03:42:58.715967 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:42:58 crc kubenswrapper[4921]: E0103 03:42:58.716075 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:42:59 crc kubenswrapper[4921]: I0103 03:42:59.883550 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:42:59 crc kubenswrapper[4921]: E0103 03:42:59.884636 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:42:59 crc kubenswrapper[4921]: I0103 03:42:59.883697 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:42:59 crc kubenswrapper[4921]: E0103 03:42:59.884838 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:42:59 crc kubenswrapper[4921]: I0103 03:42:59.883620 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:42:59 crc kubenswrapper[4921]: E0103 03:42:59.884999 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:00 crc kubenswrapper[4921]: I0103 03:43:00.883123 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:00 crc kubenswrapper[4921]: E0103 03:43:00.883786 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:01 crc kubenswrapper[4921]: I0103 03:43:01.883590 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:01 crc kubenswrapper[4921]: I0103 03:43:01.883755 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:01 crc kubenswrapper[4921]: E0103 03:43:01.884442 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:01 crc kubenswrapper[4921]: I0103 03:43:01.883785 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:01 crc kubenswrapper[4921]: E0103 03:43:01.884686 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:01 crc kubenswrapper[4921]: E0103 03:43:01.885064 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:02 crc kubenswrapper[4921]: I0103 03:43:02.883208 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:02 crc kubenswrapper[4921]: E0103 03:43:02.883468 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:03 crc kubenswrapper[4921]: I0103 03:43:03.883738 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:03 crc kubenswrapper[4921]: I0103 03:43:03.883832 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:03 crc kubenswrapper[4921]: I0103 03:43:03.883885 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:03 crc kubenswrapper[4921]: E0103 03:43:03.884085 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:03 crc kubenswrapper[4921]: E0103 03:43:03.884194 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:03 crc kubenswrapper[4921]: E0103 03:43:03.884365 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:04 crc kubenswrapper[4921]: I0103 03:43:04.883403 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:04 crc kubenswrapper[4921]: E0103 03:43:04.884302 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:04 crc kubenswrapper[4921]: E0103 03:43:04.901687 4921 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 03 03:43:04 crc kubenswrapper[4921]: E0103 03:43:04.980215 4921 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 03 03:43:05 crc kubenswrapper[4921]: I0103 03:43:05.883540 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:05 crc kubenswrapper[4921]: E0103 03:43:05.884181 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:05 crc kubenswrapper[4921]: I0103 03:43:05.883768 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:05 crc kubenswrapper[4921]: E0103 03:43:05.884490 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:05 crc kubenswrapper[4921]: I0103 03:43:05.883593 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:05 crc kubenswrapper[4921]: E0103 03:43:05.885251 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:06 crc kubenswrapper[4921]: I0103 03:43:06.883199 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:06 crc kubenswrapper[4921]: E0103 03:43:06.883495 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:07 crc kubenswrapper[4921]: I0103 03:43:07.883145 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:07 crc kubenswrapper[4921]: I0103 03:43:07.883256 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:07 crc kubenswrapper[4921]: E0103 03:43:07.883329 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:07 crc kubenswrapper[4921]: E0103 03:43:07.883511 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:07 crc kubenswrapper[4921]: I0103 03:43:07.883538 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:07 crc kubenswrapper[4921]: E0103 03:43:07.883640 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:08 crc kubenswrapper[4921]: I0103 03:43:08.883902 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:08 crc kubenswrapper[4921]: E0103 03:43:08.884161 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:09 crc kubenswrapper[4921]: I0103 03:43:09.883864 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:09 crc kubenswrapper[4921]: I0103 03:43:09.883885 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:09 crc kubenswrapper[4921]: E0103 03:43:09.884219 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:09 crc kubenswrapper[4921]: I0103 03:43:09.884372 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:09 crc kubenswrapper[4921]: I0103 03:43:09.884449 4921 scope.go:117] "RemoveContainer" containerID="5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b" Jan 03 03:43:09 crc kubenswrapper[4921]: E0103 03:43:09.884523 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:09 crc kubenswrapper[4921]: E0103 03:43:09.884701 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:09 crc kubenswrapper[4921]: E0103 03:43:09.981170 4921 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 03 03:43:10 crc kubenswrapper[4921]: I0103 03:43:10.631523 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/1.log" Jan 03 03:43:10 crc kubenswrapper[4921]: I0103 03:43:10.631590 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerStarted","Data":"956614e9298ecf848c23f3839a090463fa6a357c9bf13d8ba23ff555363d03cf"} Jan 03 03:43:10 crc kubenswrapper[4921]: I0103 03:43:10.883077 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:10 crc kubenswrapper[4921]: E0103 03:43:10.883311 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:11 crc kubenswrapper[4921]: I0103 03:43:11.883262 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:11 crc kubenswrapper[4921]: I0103 03:43:11.883316 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:11 crc kubenswrapper[4921]: I0103 03:43:11.883261 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:11 crc kubenswrapper[4921]: E0103 03:43:11.883455 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:11 crc kubenswrapper[4921]: E0103 03:43:11.883655 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:11 crc kubenswrapper[4921]: E0103 03:43:11.883777 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:12 crc kubenswrapper[4921]: I0103 03:43:12.883163 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:12 crc kubenswrapper[4921]: E0103 03:43:12.883402 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:13 crc kubenswrapper[4921]: I0103 03:43:13.882769 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:13 crc kubenswrapper[4921]: I0103 03:43:13.882820 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:13 crc kubenswrapper[4921]: I0103 03:43:13.882923 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:13 crc kubenswrapper[4921]: E0103 03:43:13.882977 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 03 03:43:13 crc kubenswrapper[4921]: E0103 03:43:13.883189 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 03 03:43:13 crc kubenswrapper[4921]: E0103 03:43:13.883474 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 03 03:43:14 crc kubenswrapper[4921]: I0103 03:43:14.883233 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:14 crc kubenswrapper[4921]: E0103 03:43:14.885748 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vsp88" podUID="fb5f46fd-64d0-4308-8fc6-5eff70d2521b" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.883513 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.883594 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.883594 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.886264 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.888858 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.888945 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 03 03:43:15 crc kubenswrapper[4921]: I0103 03:43:15.890642 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 03 03:43:16 crc kubenswrapper[4921]: I0103 03:43:16.883633 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:16 crc kubenswrapper[4921]: I0103 03:43:16.887488 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 03 03:43:16 crc kubenswrapper[4921]: I0103 03:43:16.887807 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 03 03:43:17 crc kubenswrapper[4921]: I0103 03:43:17.359708 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.429816 4921 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.481300 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2dn4r"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.481789 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.485386 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.485554 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.485755 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.485865 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.485870 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.486062 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.486158 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.486908 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.487451 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.487539 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.496640 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.497324 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.498222 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.498634 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.498776 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.498879 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.499460 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.500092 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.500230 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.500469 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.500630 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.500830 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.501448 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.501484 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.501606 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.501681 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.501818 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.502402 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.502723 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.510380 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qmx7w"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.511118 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.511215 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5vdkd"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.511958 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.512012 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.512758 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.513429 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62z7t"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.514052 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6psdt"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.514459 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.514629 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.520748 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-gqd64"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.522514 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.545794 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.546663 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.547332 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.556423 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.556880 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.558815 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2855s"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.558954 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.560172 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qr9zc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.560308 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575088 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575130 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-client\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575173 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g2zf\" (UniqueName: \"kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575195 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575217 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxxtz\" (UniqueName: \"kubernetes.io/projected/2336e873-d98d-4dc4-81f0-9327afe9d200-kube-api-access-qxxtz\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575237 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575254 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575285 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-serving-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575303 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-encryption-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575320 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-images\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575339 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575357 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqps7\" (UniqueName: \"kubernetes.io/projected/51c2ba01-1111-4f93-af2f-a6d37f756ace-kube-api-access-pqps7\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575374 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j47d\" (UniqueName: \"kubernetes.io/projected/478cd3d6-1d5e-4fe0-a103-2caa602ed743-kube-api-access-7j47d\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575389 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-image-import-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575413 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkgtv\" (UniqueName: \"kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575431 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-node-pullsecrets\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575450 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575466 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-serving-cert\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575482 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-service-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575498 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575515 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575531 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnt2f\" (UniqueName: \"kubernetes.io/projected/54a51409-4f6d-4c24-9242-a4274fe755a7-kube-api-access-fnt2f\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575548 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575562 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-service-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575588 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-client\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575608 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-config\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575627 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-config\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575645 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skwmr\" (UniqueName: \"kubernetes.io/projected/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-kube-api-access-skwmr\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575659 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-audit-dir\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575674 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-trusted-ca\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575688 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575705 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-config\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575721 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575735 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjnjd\" (UniqueName: \"kubernetes.io/projected/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-kube-api-access-bjnjd\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575753 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-encryption-config\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575771 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-serving-cert\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575786 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-dir\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575802 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-audit\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575818 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575835 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-config\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575857 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-serving-cert\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575877 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-serving-cert\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575897 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjhb2\" (UniqueName: \"kubernetes.io/projected/ab4a6d77-7628-46ae-9304-da7a2149c336-kube-api-access-cjhb2\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575914 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575933 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-policies\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575951 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bhbq\" (UniqueName: \"kubernetes.io/projected/9fa8138b-e464-4c9e-8db6-09e0f03e8507-kube-api-access-8bhbq\") pod \"downloads-7954f5f757-gqd64\" (UID: \"9fa8138b-e464-4c9e-8db6-09e0f03e8507\") " pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575966 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2336e873-d98d-4dc4-81f0-9327afe9d200-serving-cert\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.575987 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.576005 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.576027 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-client\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.576053 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/478cd3d6-1d5e-4fe0-a103-2caa602ed743-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.577002 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.577356 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.577856 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.577965 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.578429 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.578556 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.578858 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.580044 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.580644 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.595007 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4sw9z"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.595831 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.596403 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-x8wkt"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.596875 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.597106 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.597107 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.598011 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.598732 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.599071 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.599251 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.600286 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.601066 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.601364 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.601755 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.601820 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.602188 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.602658 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.602677 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.603225 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.614095 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.614411 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.614848 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.615171 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.615332 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.619863 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.624745 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.625036 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.625643 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.625961 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.627477 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.627996 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.628203 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.628220 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.629840 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2dn4r"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.629917 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lsjpr"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.630781 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.631355 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632323 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632466 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632575 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632834 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632843 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632893 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.632991 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633126 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633319 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633432 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633603 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633812 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633835 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633982 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633994 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634027 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634101 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634112 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633989 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634188 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634195 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634309 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634333 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634380 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634423 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634461 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634517 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634552 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634598 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634617 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634692 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634742 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634764 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634839 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634847 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634863 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635100 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635170 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635243 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634116 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633604 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635331 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634936 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634968 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.634995 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635023 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633612 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635051 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.633445 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635084 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.635737 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.636204 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.636354 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.636398 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.640078 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.640197 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.640353 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.640467 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.641086 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.641297 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.641538 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.642194 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.670762 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9j876"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.677325 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.677870 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.678602 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.679355 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.681738 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684704 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqps7\" (UniqueName: \"kubernetes.io/projected/51c2ba01-1111-4f93-af2f-a6d37f756ace-kube-api-access-pqps7\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684747 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j47d\" (UniqueName: \"kubernetes.io/projected/478cd3d6-1d5e-4fe0-a103-2caa602ed743-kube-api-access-7j47d\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684794 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-image-import-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684886 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkgtv\" (UniqueName: \"kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684924 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-node-pullsecrets\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684948 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.684969 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-serving-cert\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685116 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-service-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685142 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685161 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-service-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685318 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685340 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnt2f\" (UniqueName: \"kubernetes.io/projected/54a51409-4f6d-4c24-9242-a4274fe755a7-kube-api-access-fnt2f\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685362 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685457 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-client\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685491 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-config\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685512 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-config\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685530 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skwmr\" (UniqueName: \"kubernetes.io/projected/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-kube-api-access-skwmr\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685659 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-audit-dir\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685686 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685714 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-trusted-ca\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685946 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-config\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.685979 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686003 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjnjd\" (UniqueName: \"kubernetes.io/projected/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-kube-api-access-bjnjd\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686132 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-encryption-config\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686173 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-serving-cert\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686190 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-dir\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686221 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-audit\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686248 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686430 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-serving-cert\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.686450 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-config\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.687569 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.693794 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.702503 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-dir\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.702643 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-config\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.703220 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.703252 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.703561 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-serving-cert\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.703722 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjhb2\" (UniqueName: \"kubernetes.io/projected/ab4a6d77-7628-46ae-9304-da7a2149c336-kube-api-access-cjhb2\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.687053 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.711886 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.713977 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.714235 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-policies\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.714469 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bhbq\" (UniqueName: \"kubernetes.io/projected/9fa8138b-e464-4c9e-8db6-09e0f03e8507-kube-api-access-8bhbq\") pod \"downloads-7954f5f757-gqd64\" (UID: \"9fa8138b-e464-4c9e-8db6-09e0f03e8507\") " pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.714627 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2336e873-d98d-4dc4-81f0-9327afe9d200-serving-cert\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.716514 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.716731 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-client\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.716987 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-config\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717283 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-serving-cert\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717304 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-audit-policies\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717464 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717485 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717503 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/478cd3d6-1d5e-4fe0-a103-2caa602ed743-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717542 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717561 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-client\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717655 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g2zf\" (UniqueName: \"kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717673 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717691 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxxtz\" (UniqueName: \"kubernetes.io/projected/2336e873-d98d-4dc4-81f0-9327afe9d200-kube-api-access-qxxtz\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717713 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-serving-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717744 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717760 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717776 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-encryption-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.717794 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-images\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.719560 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.725295 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-serving-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.728474 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.730642 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-config\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.733420 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.735147 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.735379 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.735630 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.735897 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.736906 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-service-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.736957 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-image-import-ca\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.737031 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.737720 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-audit-dir\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.738559 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.738712 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ab4a6d77-7628-46ae-9304-da7a2149c336-node-pullsecrets\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.738848 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.739057 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-audit\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.739359 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-encryption-config\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.739652 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/478cd3d6-1d5e-4fe0-a103-2caa602ed743-images\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.740044 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-config\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.740404 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.740700 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.740869 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-serving-cert\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.744103 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.744422 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.746693 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-client\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.746946 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-service-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.747948 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-serving-cert\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.748380 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.748589 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.748668 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.748856 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.749386 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.750879 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2336e873-d98d-4dc4-81f0-9327afe9d200-serving-cert\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.750947 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.751251 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-etcd-client\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.751743 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/54a51409-4f6d-4c24-9242-a4274fe755a7-encryption-config\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.751823 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/478cd3d6-1d5e-4fe0-a103-2caa602ed743-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.752198 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/51c2ba01-1111-4f93-af2f-a6d37f756ace-etcd-ca\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.753183 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab4a6d77-7628-46ae-9304-da7a2149c336-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.757917 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.759701 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-etcd-client\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.761587 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.764402 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.764605 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.764802 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.765075 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.766651 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.766650 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.767067 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab4a6d77-7628-46ae-9304-da7a2149c336-serving-cert\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.767147 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.768342 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qmx7w"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.769179 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2336e873-d98d-4dc4-81f0-9327afe9d200-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.769605 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-9tl8f"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.770572 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.770913 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-trusted-ca\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.771126 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k4ffk"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.780185 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.780816 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.780902 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.783918 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.784057 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.783930 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5vdkd"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.790209 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62z7t"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.792781 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.797658 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gqd64"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.798890 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2855s"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.800503 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6psdt"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.802134 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qr9zc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.803648 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.804767 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.806893 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.808444 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.809977 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.811435 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.812779 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.814500 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.815728 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.817191 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.818729 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.820228 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.821728 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.822822 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.823971 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4sw9z"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.825180 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-kw6ql"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.825335 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.826375 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.827792 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.828949 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.830053 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.831258 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lsjpr"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.832291 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9j876"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.833401 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.834490 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.835604 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.836745 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.838050 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kw6ql"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.838833 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k4ffk"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839544 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839584 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-webhook-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839607 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839689 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839733 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2fc9836-7783-4f2d-82e3-4750d96b5f07-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839766 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-oauth-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839791 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839812 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9kgj\" (UniqueName: \"kubernetes.io/projected/5697e250-01f6-4047-8399-ca18e3c17cd5-kube-api-access-h9kgj\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839839 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-default-certificate\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839863 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839881 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6k5\" (UniqueName: \"kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839905 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f618317d-4a5b-465b-b14c-72ae7ed50c9c-tmpfs\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839923 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839959 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.839982 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-service-ca-bundle\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840003 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840027 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840034 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840254 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks8qt\" (UniqueName: \"kubernetes.io/projected/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-kube-api-access-ks8qt\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840351 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqjtx\" (UniqueName: \"kubernetes.io/projected/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-kube-api-access-zqjtx\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840404 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840425 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5697e250-01f6-4047-8399-ca18e3c17cd5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840446 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840507 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-proxy-tls\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840538 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840641 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lvsd\" (UniqueName: \"kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840675 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-stats-auth\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840697 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqcbx\" (UniqueName: \"kubernetes.io/projected/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-kube-api-access-fqcbx\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840721 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840740 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2fc9836-7783-4f2d-82e3-4750d96b5f07-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840758 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840777 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htcwp\" (UniqueName: \"kubernetes.io/projected/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-kube-api-access-htcwp\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840798 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-metrics-certs\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840820 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm2mx\" (UniqueName: \"kubernetes.io/projected/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-kube-api-access-lm2mx\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840838 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-trusted-ca-bundle\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840859 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840901 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-service-ca\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840920 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-oauth-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840944 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840965 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkrjh\" (UniqueName: \"kubernetes.io/projected/393f5b19-1c3d-4521-8cda-002f27f95734-kube-api-access-rkrjh\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.840984 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp9mq\" (UniqueName: \"kubernetes.io/projected/a2fc9836-7783-4f2d-82e3-4750d96b5f07-kube-api-access-fp9mq\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841011 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841034 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841057 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kskf\" (UniqueName: \"kubernetes.io/projected/f618317d-4a5b-465b-b14c-72ae7ed50c9c-kube-api-access-8kskf\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841078 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841095 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-apiservice-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841114 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841133 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841170 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841192 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwsjp\" (UniqueName: \"kubernetes.io/projected/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-kube-api-access-fwsjp\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841215 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.841235 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9tl8f"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.842217 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-sbgkw"] Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.843095 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.844202 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.864593 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.884022 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.904632 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.925815 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942574 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942631 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kskf\" (UniqueName: \"kubernetes.io/projected/f618317d-4a5b-465b-b14c-72ae7ed50c9c-kube-api-access-8kskf\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942660 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942680 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-apiservice-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942724 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942747 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942794 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942817 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwsjp\" (UniqueName: \"kubernetes.io/projected/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-kube-api-access-fwsjp\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942835 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942874 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942891 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-webhook-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942909 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942962 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.942999 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2fc9836-7783-4f2d-82e3-4750d96b5f07-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943063 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-oauth-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943085 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943104 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9kgj\" (UniqueName: \"kubernetes.io/projected/5697e250-01f6-4047-8399-ca18e3c17cd5-kube-api-access-h9kgj\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943152 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-default-certificate\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943173 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943194 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6k5\" (UniqueName: \"kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943217 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f618317d-4a5b-465b-b14c-72ae7ed50c9c-tmpfs\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943236 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943283 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943307 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-service-ca-bundle\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943330 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943349 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943388 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks8qt\" (UniqueName: \"kubernetes.io/projected/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-kube-api-access-ks8qt\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943405 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqjtx\" (UniqueName: \"kubernetes.io/projected/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-kube-api-access-zqjtx\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943423 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943440 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5697e250-01f6-4047-8399-ca18e3c17cd5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943460 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943480 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-proxy-tls\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943499 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943548 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lvsd\" (UniqueName: \"kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943566 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-stats-auth\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943585 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqcbx\" (UniqueName: \"kubernetes.io/projected/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-kube-api-access-fqcbx\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943603 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943620 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2fc9836-7783-4f2d-82e3-4750d96b5f07-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943639 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943656 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htcwp\" (UniqueName: \"kubernetes.io/projected/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-kube-api-access-htcwp\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943673 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-metrics-certs\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943691 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm2mx\" (UniqueName: \"kubernetes.io/projected/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-kube-api-access-lm2mx\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943707 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-trusted-ca-bundle\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943723 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943761 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-service-ca\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943778 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-oauth-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943799 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943816 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkrjh\" (UniqueName: \"kubernetes.io/projected/393f5b19-1c3d-4521-8cda-002f27f95734-kube-api-access-rkrjh\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943866 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp9mq\" (UniqueName: \"kubernetes.io/projected/a2fc9836-7783-4f2d-82e3-4750d96b5f07-kube-api-access-fp9mq\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.943951 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.945792 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.946710 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.946929 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f618317d-4a5b-465b-b14c-72ae7ed50c9c-tmpfs\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.947706 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.947818 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.948552 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.948710 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-service-ca\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.950080 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.950480 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.951671 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a2fc9836-7783-4f2d-82e3-4750d96b5f07-available-featuregates\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.952347 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.952539 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-trusted-ca-bundle\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.952876 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.953053 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.953473 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-oauth-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.953757 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.953834 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.954532 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.954747 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.956072 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.963472 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-stats-auth\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.965175 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.965313 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.965458 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-oauth-config\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.965800 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2fc9836-7783-4f2d-82e3-4750d96b5f07-serving-cert\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.966092 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.967342 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-console-serving-cert\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.967976 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.983985 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 03 03:43:20 crc kubenswrapper[4921]: I0103 03:43:20.988817 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-service-ca-bundle\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.004884 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.014594 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-metrics-certs\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.024040 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.044442 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.057336 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-default-certificate\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.064593 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.084823 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.105355 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.124544 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.145180 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.164181 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.184721 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.205481 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.224749 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.245692 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.264181 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.285924 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.305230 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.325841 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.344705 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.350157 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-webhook-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.351721 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f618317d-4a5b-465b-b14c-72ae7ed50c9c-apiservice-cert\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.364746 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.384939 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.404338 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.425313 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.445352 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.464927 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.485122 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.506951 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.524848 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.547111 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.565236 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.580743 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-proxy-tls\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.584756 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.605307 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.623028 4921 request.go:700] Waited for 1.007457935s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.625151 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.644528 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.665715 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.684306 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.704450 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.725248 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.745078 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.764189 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.785029 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.815157 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.825540 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.825673 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.846425 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.885988 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.896975 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5697e250-01f6-4047-8399-ca18e3c17cd5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.904881 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.917220 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.925111 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.943093 4921 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.943437 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key podName:393f5b19-1c3d-4521-8cda-002f27f95734 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:22.443386438 +0000 UTC m=+138.054813262 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key") pod "service-ca-9c57cc56f-9j876" (UID: "393f5b19-1c3d-4521-8cda-002f27f95734") : failed to sync secret cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.945406 4921 secret.go:188] Couldn't get secret openshift-dns-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.945517 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls podName:ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe nodeName:}" failed. No retries permitted until 2026-01-03 03:43:22.445486995 +0000 UTC m=+138.056913879 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls") pod "dns-operator-744455d44c-lsjpr" (UID: "ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe") : failed to sync secret cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.946020 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.953880 4921 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: E0103 03:43:21.953996 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle podName:393f5b19-1c3d-4521-8cda-002f27f95734 nodeName:}" failed. No retries permitted until 2026-01-03 03:43:22.453973002 +0000 UTC m=+138.065399836 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle") pod "service-ca-9c57cc56f-9j876" (UID: "393f5b19-1c3d-4521-8cda-002f27f95734") : failed to sync configmap cache: timed out waiting for the condition Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.965509 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 03 03:43:21 crc kubenswrapper[4921]: I0103 03:43:21.985536 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.004904 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.025174 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.046767 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.066079 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.086128 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.109937 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.140146 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j47d\" (UniqueName: \"kubernetes.io/projected/478cd3d6-1d5e-4fe0-a103-2caa602ed743-kube-api-access-7j47d\") pod \"machine-api-operator-5694c8668f-2dn4r\" (UID: \"478cd3d6-1d5e-4fe0-a103-2caa602ed743\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.163633 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.170212 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqps7\" (UniqueName: \"kubernetes.io/projected/51c2ba01-1111-4f93-af2f-a6d37f756ace-kube-api-access-pqps7\") pod \"etcd-operator-b45778765-6psdt\" (UID: \"51c2ba01-1111-4f93-af2f-a6d37f756ace\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.185909 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.204156 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.243787 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjhb2\" (UniqueName: \"kubernetes.io/projected/ab4a6d77-7628-46ae-9304-da7a2149c336-kube-api-access-cjhb2\") pod \"apiserver-76f77b778f-5vdkd\" (UID: \"ab4a6d77-7628-46ae-9304-da7a2149c336\") " pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.243999 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.284689 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bhbq\" (UniqueName: \"kubernetes.io/projected/9fa8138b-e464-4c9e-8db6-09e0f03e8507-kube-api-access-8bhbq\") pod \"downloads-7954f5f757-gqd64\" (UID: \"9fa8138b-e464-4c9e-8db6-09e0f03e8507\") " pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.284969 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.296031 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.305290 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.342065 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjnjd\" (UniqueName: \"kubernetes.io/projected/66b61c47-d1cd-4c4c-b335-8e20bfd70f5a-kube-api-access-bjnjd\") pod \"openshift-apiserver-operator-796bbdcf4f-4hxgb\" (UID: \"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.365916 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skwmr\" (UniqueName: \"kubernetes.io/projected/c4b61ff0-15aa-4a89-8bf3-e4d6dda31770-kube-api-access-skwmr\") pod \"console-operator-58897d9998-62z7t\" (UID: \"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770\") " pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.383827 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxxtz\" (UniqueName: \"kubernetes.io/projected/2336e873-d98d-4dc4-81f0-9327afe9d200-kube-api-access-qxxtz\") pod \"authentication-operator-69f744f599-qmx7w\" (UID: \"2336e873-d98d-4dc4-81f0-9327afe9d200\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.391386 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.401663 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnt2f\" (UniqueName: \"kubernetes.io/projected/54a51409-4f6d-4c24-9242-a4274fe755a7-kube-api-access-fnt2f\") pod \"apiserver-7bbb656c7d-znd66\" (UID: \"54a51409-4f6d-4c24-9242-a4274fe755a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.428607 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.430054 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkgtv\" (UniqueName: \"kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv\") pod \"controller-manager-879f6c89f-xvdtc\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.438740 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.443072 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g2zf\" (UniqueName: \"kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf\") pod \"route-controller-manager-6576b87f9c-cwfsp\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.445552 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.465889 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.482585 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.482685 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.482810 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.483090 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.484817 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/393f5b19-1c3d-4521-8cda-002f27f95734-signing-cabundle\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.486520 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.486983 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/393f5b19-1c3d-4521-8cda-002f27f95734-signing-key\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.487639 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-metrics-tls\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.505618 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.524743 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.544917 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gqd64"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.545115 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 03 03:43:22 crc kubenswrapper[4921]: W0103 03:43:22.555919 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fa8138b_e464_4c9e_8db6_09e0f03e8507.slice/crio-9f4a09d618d1f8a9cbe8af1f60742bda9ec4a438f8c94aae6ac30977bd4ade77 WatchSource:0}: Error finding container 9f4a09d618d1f8a9cbe8af1f60742bda9ec4a438f8c94aae6ac30977bd4ade77: Status 404 returned error can't find the container with id 9f4a09d618d1f8a9cbe8af1f60742bda9ec4a438f8c94aae6ac30977bd4ade77 Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.565757 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.585475 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.586719 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.605116 4921 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.624376 4921 request.go:700] Waited for 1.797677293s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-dockercfg-jwfmh&limit=500&resourceVersion=0 Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.624385 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.627221 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.644622 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.645704 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-2dn4r"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.657144 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.664591 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.682985 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.684312 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.705996 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.729302 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.739057 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" event={"ID":"478cd3d6-1d5e-4fe0-a103-2caa602ed743","Type":"ContainerStarted","Data":"86b5c14be720e24e01b9cac617cdde6c4d600c88ab1f029cfe526897b4a2588f"} Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.745385 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.745455 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gqd64" event={"ID":"9fa8138b-e464-4c9e-8db6-09e0f03e8507","Type":"ContainerStarted","Data":"9f4a09d618d1f8a9cbe8af1f60742bda9ec4a438f8c94aae6ac30977bd4ade77"} Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.783347 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kskf\" (UniqueName: \"kubernetes.io/projected/f618317d-4a5b-465b-b14c-72ae7ed50c9c-kube-api-access-8kskf\") pod \"packageserver-d55dfcdfc-m7q7x\" (UID: \"f618317d-4a5b-465b-b14c-72ae7ed50c9c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.788953 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwsjp\" (UniqueName: \"kubernetes.io/projected/ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe-kube-api-access-fwsjp\") pod \"dns-operator-744455d44c-lsjpr\" (UID: \"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe\") " pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.802584 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htcwp\" (UniqueName: \"kubernetes.io/projected/8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0-kube-api-access-htcwp\") pod \"control-plane-machine-set-operator-78cbb6b69f-fxkcl\" (UID: \"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.818767 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lvsd\" (UniqueName: \"kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd\") pod \"marketplace-operator-79b997595-zng4m\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.836351 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.843203 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks8qt\" (UniqueName: \"kubernetes.io/projected/13f1ddbe-b27a-4db0-bd05-a37b298bdebd-kube-api-access-ks8qt\") pod \"router-default-5444994796-x8wkt\" (UID: \"13f1ddbe-b27a-4db0-bd05-a37b298bdebd\") " pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.860916 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm2mx\" (UniqueName: \"kubernetes.io/projected/714b3568-8ca6-4d1a-b451-1dae7b24fbdc-kube-api-access-lm2mx\") pod \"machine-config-controller-84d6567774-96kc8\" (UID: \"714b3568-8ca6-4d1a-b451-1dae7b24fbdc\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.861525 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62z7t"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.890026 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqjtx\" (UniqueName: \"kubernetes.io/projected/2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9-kube-api-access-zqjtx\") pod \"console-f9d7485db-2855s\" (UID: \"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9\") " pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.904257 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.904310 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5vdkd"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.904323 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6psdt"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.905260 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qmx7w"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.908252 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkrjh\" (UniqueName: \"kubernetes.io/projected/393f5b19-1c3d-4521-8cda-002f27f95734-kube-api-access-rkrjh\") pod \"service-ca-9c57cc56f-9j876\" (UID: \"393f5b19-1c3d-4521-8cda-002f27f95734\") " pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.908658 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.920663 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqcbx\" (UniqueName: \"kubernetes.io/projected/a2245ae0-ac63-4503-bbd5-59841ca0e9a6-kube-api-access-fqcbx\") pod \"cluster-samples-operator-665b6dd947-8wjxs\" (UID: \"a2245ae0-ac63-4503-bbd5-59841ca0e9a6\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.921566 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.935245 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.937124 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp9mq\" (UniqueName: \"kubernetes.io/projected/a2fc9836-7783-4f2d-82e3-4750d96b5f07-kube-api-access-fp9mq\") pod \"openshift-config-operator-7777fb866f-hsm7c\" (UID: \"a2fc9836-7783-4f2d-82e3-4750d96b5f07\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.956447 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6k5\" (UniqueName: \"kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5\") pod \"oauth-openshift-558db77b4-qr9zc\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.965443 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.976384 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9kgj\" (UniqueName: \"kubernetes.io/projected/5697e250-01f6-4047-8399-ca18e3c17cd5-kube-api-access-h9kgj\") pod \"package-server-manager-789f6589d5-zzrtr\" (UID: \"5697e250-01f6-4047-8399-ca18e3c17cd5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:22 crc kubenswrapper[4921]: I0103 03:43:22.982112 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.005540 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.616651 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.623300 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9j876" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.623819 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.624062 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.624524 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.624771 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.627017 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.630883 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.631054 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.631564 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.634402 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.634527 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.634581 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr6cz\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.634627 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.634655 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: E0103 03:43:23.635751 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.135710156 +0000 UTC m=+139.747137020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:23 crc kubenswrapper[4921]: W0103 03:43:23.658573 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d74e762_4edc_473e_b643_af99acf8f210.slice/crio-22b79e7c6e242bff4c37835d425cc6cc0c4e2f009c7a7c3009db1557ae867f79 WatchSource:0}: Error finding container 22b79e7c6e242bff4c37835d425cc6cc0c4e2f009c7a7c3009db1557ae867f79: Status 404 returned error can't find the container with id 22b79e7c6e242bff4c37835d425cc6cc0c4e2f009c7a7c3009db1557ae867f79 Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.736885 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737073 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f36c145-b6cf-42e1-be98-11521fa8d1f9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737112 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16503625-6d0f-42a8-b7fb-991ae13be2d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737175 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-srv-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737203 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737227 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-socket-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737252 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7bbd\" (UniqueName: \"kubernetes.io/projected/50b7d4a0-868d-4ce3-a193-974ad0ef8837-kube-api-access-l7bbd\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737299 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-profile-collector-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737321 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-plugins-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737368 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv6ft\" (UniqueName: \"kubernetes.io/projected/8fa84952-f5e1-4a40-b263-31b8d6a8e361-kube-api-access-fv6ft\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737421 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16503625-6d0f-42a8-b7fb-991ae13be2d4-config\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737525 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737559 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737584 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737607 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqc4s\" (UniqueName: \"kubernetes.io/projected/080d2cc3-9d0f-4c08-a279-ab67b3877635-kube-api-access-pqc4s\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737644 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737689 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27s2s\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-kube-api-access-27s2s\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737714 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-node-bootstrap-token\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737736 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737794 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2wd4\" (UniqueName: \"kubernetes.io/projected/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-kube-api-access-x2wd4\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737815 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twqjw\" (UniqueName: \"kubernetes.io/projected/c2e8f120-d9bf-4f9b-960e-5f816329d194-kube-api-access-twqjw\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737845 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-registration-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737872 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr6cz\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737895 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgd6z\" (UniqueName: \"kubernetes.io/projected/dc0605d6-fd4e-41cd-b8f6-27a304c252c0-kube-api-access-dgd6z\") pod \"migrator-59844c95c7-jvxlw\" (UID: \"dc0605d6-fd4e-41cd-b8f6-27a304c252c0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737920 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f35ae7-e556-4393-bb87-eb9601eb36bb-proxy-tls\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737944 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df1cb7b9-01cf-4b52-9556-3f84680dd057-trusted-ca\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737969 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.737990 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f36c145-b6cf-42e1-be98-11521fa8d1f9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738011 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-config-volume\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738037 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738063 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-certs\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738139 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738163 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58lg2\" (UniqueName: \"kubernetes.io/projected/59f35ae7-e556-4393-bb87-eb9601eb36bb-kube-api-access-58lg2\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738185 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f625709-e003-4472-990b-7d7bb1c81477-machine-approver-tls\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738222 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-metrics-tls\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.738245 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64578484-1e9f-49cd-808b-8f036bb9bd48-serving-cert\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: E0103 03:43:23.738476 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.238448958 +0000 UTC m=+139.849875792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.740297 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746173 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e89c0-7bb3-4e34-a770-265dfc317628-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746240 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33e89c0-7bb3-4e34-a770-265dfc317628-config\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746312 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-auth-proxy-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746342 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e8f120-d9bf-4f9b-960e-5f816329d194-cert\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746371 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746396 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9c5q\" (UniqueName: \"kubernetes.io/projected/64578484-1e9f-49cd-808b-8f036bb9bd48-kube-api-access-b9c5q\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746452 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55705621-48c6-4662-b10c-492b16cb1368-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746509 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746544 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flfs7\" (UniqueName: \"kubernetes.io/projected/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-kube-api-access-flfs7\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746596 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746070 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.746629 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55705621-48c6-4662-b10c-492b16cb1368-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749803 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749870 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749887 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-srv-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749938 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/df1cb7b9-01cf-4b52-9556-3f84680dd057-metrics-tls\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749958 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j95dg\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-kube-api-access-j95dg\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.749975 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/080d2cc3-9d0f-4c08-a279-ab67b3877635-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750027 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55705621-48c6-4662-b10c-492b16cb1368-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750044 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nfsp\" (UniqueName: \"kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750065 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnck8\" (UniqueName: \"kubernetes.io/projected/7f36c145-b6cf-42e1-be98-11521fa8d1f9-kube-api-access-jnck8\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750099 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16503625-6d0f-42a8-b7fb-991ae13be2d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750197 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750301 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p82jv\" (UniqueName: \"kubernetes.io/projected/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-kube-api-access-p82jv\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750323 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74kv4\" (UniqueName: \"kubernetes.io/projected/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-kube-api-access-74kv4\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750364 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750384 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750401 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c33e89c0-7bb3-4e34-a770-265dfc317628-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750439 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtwb7\" (UniqueName: \"kubernetes.io/projected/8f625709-e003-4472-990b-7d7bb1c81477-kube-api-access-dtwb7\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750466 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-images\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750485 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64578484-1e9f-49cd-808b-8f036bb9bd48-config\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750556 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-csi-data-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.750576 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-mountpoint-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.754623 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.756833 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.762796 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.767046 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr6cz\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.767736 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.778392 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" event={"ID":"ab4a6d77-7628-46ae-9304-da7a2149c336","Type":"ContainerStarted","Data":"7a74d4440f4fb4b7f39990930fdee9df53c4f5de46b19873fe59c8d89c6d79bf"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.789249 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" event={"ID":"2336e873-d98d-4dc4-81f0-9327afe9d200","Type":"ContainerStarted","Data":"a2ca973e7da4f07d4adab06f1c5ec29c62aeb3cc695e951a46825fcee2a94be2"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.790830 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" event={"ID":"54a51409-4f6d-4c24-9242-a4274fe755a7","Type":"ContainerStarted","Data":"6518f813351a2da4811fb1b2a6d1197d849e4a0a9d74787ed9791d124333a661"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.792335 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gqd64" event={"ID":"9fa8138b-e464-4c9e-8db6-09e0f03e8507","Type":"ContainerStarted","Data":"bfc146dbaea0b1f17b907c8f5ace1b3e29ed60b14302e21d098882e0f63dfe61"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.793261 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.794590 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" event={"ID":"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a","Type":"ContainerStarted","Data":"06cd937127776e40553c024cdcf103a4b10cafd2a478e4b3ff9136b863072aa5"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.795664 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-62z7t" event={"ID":"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770","Type":"ContainerStarted","Data":"1f5cb7294b5b33c0158517b8e76499e7fbbd2b68db03d4f92d1697c2e3b97be2"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.797515 4921 patch_prober.go:28] interesting pod/downloads-7954f5f757-gqd64 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.797565 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gqd64" podUID="9fa8138b-e464-4c9e-8db6-09e0f03e8507" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.806142 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" event={"ID":"1d74e762-4edc-473e-b643-af99acf8f210","Type":"ContainerStarted","Data":"22b79e7c6e242bff4c37835d425cc6cc0c4e2f009c7a7c3009db1557ae867f79"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.807883 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" event={"ID":"478cd3d6-1d5e-4fe0-a103-2caa602ed743","Type":"ContainerStarted","Data":"3ca23bbc9b97caa946c7f7661869cbf6ee2c4225abae7a753d028bdd11f0ed7e"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.810302 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" event={"ID":"4c87f43b-b9e0-4c92-985e-3dbad79ff859","Type":"ContainerStarted","Data":"b92e0bb01c864cde810ce32afea76fb7f5c1f3edd6e46d6ab4babb7b1efedb5e"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.811046 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" event={"ID":"51c2ba01-1111-4f93-af2f-a6d37f756ace","Type":"ContainerStarted","Data":"80ce34597deecbc31e4044ced0a951578f5bc6d93db9b353525e988049616bba"} Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852154 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flfs7\" (UniqueName: \"kubernetes.io/projected/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-kube-api-access-flfs7\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852197 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55705621-48c6-4662-b10c-492b16cb1368-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852221 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852259 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852305 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-srv-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852325 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/df1cb7b9-01cf-4b52-9556-3f84680dd057-metrics-tls\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852344 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/080d2cc3-9d0f-4c08-a279-ab67b3877635-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852396 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j95dg\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-kube-api-access-j95dg\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852425 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55705621-48c6-4662-b10c-492b16cb1368-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852453 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nfsp\" (UniqueName: \"kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852477 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnck8\" (UniqueName: \"kubernetes.io/projected/7f36c145-b6cf-42e1-be98-11521fa8d1f9-kube-api-access-jnck8\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852503 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16503625-6d0f-42a8-b7fb-991ae13be2d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852538 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74kv4\" (UniqueName: \"kubernetes.io/projected/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-kube-api-access-74kv4\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852557 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852574 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p82jv\" (UniqueName: \"kubernetes.io/projected/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-kube-api-access-p82jv\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852592 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c33e89c0-7bb3-4e34-a770-265dfc317628-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852615 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtwb7\" (UniqueName: \"kubernetes.io/projected/8f625709-e003-4472-990b-7d7bb1c81477-kube-api-access-dtwb7\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852634 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-images\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852652 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64578484-1e9f-49cd-808b-8f036bb9bd48-config\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852672 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-csi-data-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852692 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-mountpoint-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852711 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f36c145-b6cf-42e1-be98-11521fa8d1f9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852736 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16503625-6d0f-42a8-b7fb-991ae13be2d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852754 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-srv-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852781 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7bbd\" (UniqueName: \"kubernetes.io/projected/50b7d4a0-868d-4ce3-a193-974ad0ef8837-kube-api-access-l7bbd\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852801 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-socket-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852818 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-plugins-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852837 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-profile-collector-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852857 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv6ft\" (UniqueName: \"kubernetes.io/projected/8fa84952-f5e1-4a40-b263-31b8d6a8e361-kube-api-access-fv6ft\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852879 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16503625-6d0f-42a8-b7fb-991ae13be2d4-config\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852899 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852919 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852937 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852960 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqc4s\" (UniqueName: \"kubernetes.io/projected/080d2cc3-9d0f-4c08-a279-ab67b3877635-kube-api-access-pqc4s\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.852981 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853004 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853021 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27s2s\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-kube-api-access-27s2s\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853038 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-node-bootstrap-token\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853060 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853080 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2wd4\" (UniqueName: \"kubernetes.io/projected/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-kube-api-access-x2wd4\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853104 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-registration-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853122 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twqjw\" (UniqueName: \"kubernetes.io/projected/c2e8f120-d9bf-4f9b-960e-5f816329d194-kube-api-access-twqjw\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853140 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgd6z\" (UniqueName: \"kubernetes.io/projected/dc0605d6-fd4e-41cd-b8f6-27a304c252c0-kube-api-access-dgd6z\") pod \"migrator-59844c95c7-jvxlw\" (UID: \"dc0605d6-fd4e-41cd-b8f6-27a304c252c0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853157 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f35ae7-e556-4393-bb87-eb9601eb36bb-proxy-tls\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853176 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df1cb7b9-01cf-4b52-9556-3f84680dd057-trusted-ca\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853229 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853249 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f36c145-b6cf-42e1-be98-11521fa8d1f9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853315 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-config-volume\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853338 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-certs\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853363 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853380 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f625709-e003-4472-990b-7d7bb1c81477-machine-approver-tls\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853408 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58lg2\" (UniqueName: \"kubernetes.io/projected/59f35ae7-e556-4393-bb87-eb9601eb36bb-kube-api-access-58lg2\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853425 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64578484-1e9f-49cd-808b-8f036bb9bd48-serving-cert\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853445 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e89c0-7bb3-4e34-a770-265dfc317628-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853466 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-metrics-tls\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853486 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33e89c0-7bb3-4e34-a770-265dfc317628-config\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853510 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-auth-proxy-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853532 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853548 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e8f120-d9bf-4f9b-960e-5f816329d194-cert\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853584 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9c5q\" (UniqueName: \"kubernetes.io/projected/64578484-1e9f-49cd-808b-8f036bb9bd48-kube-api-access-b9c5q\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.853614 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55705621-48c6-4662-b10c-492b16cb1368-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.854489 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55705621-48c6-4662-b10c-492b16cb1368-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.854652 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-images\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.855201 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64578484-1e9f-49cd-808b-8f036bb9bd48-config\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.855308 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-csi-data-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.855359 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-mountpoint-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.855927 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f36c145-b6cf-42e1-be98-11521fa8d1f9-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.857820 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-socket-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.858162 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55705621-48c6-4662-b10c-492b16cb1368-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.859731 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-srv-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.860672 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: E0103 03:43:23.860908 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.360894556 +0000 UTC m=+139.972321380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.861962 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-auth-proxy-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.865593 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-metrics-tls\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.865665 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-registration-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.866438 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.867063 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c33e89c0-7bb3-4e34-a770-265dfc317628-config\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.868691 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f35ae7-e556-4393-bb87-eb9601eb36bb-proxy-tls\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.869041 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f625709-e003-4472-990b-7d7bb1c81477-config\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.869748 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-config-volume\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.871767 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-srv-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.876061 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.887186 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.887654 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64578484-1e9f-49cd-808b-8f036bb9bd48-serving-cert\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.887753 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8fa84952-f5e1-4a40-b263-31b8d6a8e361-plugins-dir\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.890304 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2e8f120-d9bf-4f9b-960e-5f816329d194-cert\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.893364 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df1cb7b9-01cf-4b52-9556-3f84680dd057-trusted-ca\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.899103 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.899851 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16503625-6d0f-42a8-b7fb-991ae13be2d4-config\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.908893 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.910440 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-node-bootstrap-token\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.912299 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f36c145-b6cf-42e1-be98-11521fa8d1f9-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.912870 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-certs\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.913162 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f35ae7-e556-4393-bb87-eb9601eb36bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.914639 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f625709-e003-4472-990b-7d7bb1c81477-machine-approver-tls\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.916644 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.917150 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/080d2cc3-9d0f-4c08-a279-ab67b3877635-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.917433 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/df1cb7b9-01cf-4b52-9556-3f84680dd057-metrics-tls\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.918788 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16503625-6d0f-42a8-b7fb-991ae13be2d4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.919235 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50b7d4a0-868d-4ce3-a193-974ad0ef8837-profile-collector-cert\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.919589 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c33e89c0-7bb3-4e34-a770-265dfc317628-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.925227 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtwb7\" (UniqueName: \"kubernetes.io/projected/8f625709-e003-4472-990b-7d7bb1c81477-kube-api-access-dtwb7\") pod \"machine-approver-56656f9798-trzp5\" (UID: \"8f625709-e003-4472-990b-7d7bb1c81477\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.925617 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgd6z\" (UniqueName: \"kubernetes.io/projected/dc0605d6-fd4e-41cd-b8f6-27a304c252c0-kube-api-access-dgd6z\") pod \"migrator-59844c95c7-jvxlw\" (UID: \"dc0605d6-fd4e-41cd-b8f6-27a304c252c0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.925734 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55705621-48c6-4662-b10c-492b16cb1368-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-zh24x\" (UID: \"55705621-48c6-4662-b10c-492b16cb1368\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.926105 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16503625-6d0f-42a8-b7fb-991ae13be2d4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-vb9dc\" (UID: \"16503625-6d0f-42a8-b7fb-991ae13be2d4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.926409 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twqjw\" (UniqueName: \"kubernetes.io/projected/c2e8f120-d9bf-4f9b-960e-5f816329d194-kube-api-access-twqjw\") pod \"ingress-canary-9tl8f\" (UID: \"c2e8f120-d9bf-4f9b-960e-5f816329d194\") " pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.927099 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27s2s\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-kube-api-access-27s2s\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.927198 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j95dg\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-kube-api-access-j95dg\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.929417 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqc4s\" (UniqueName: \"kubernetes.io/projected/080d2cc3-9d0f-4c08-a279-ab67b3877635-kube-api-access-pqc4s\") pod \"multus-admission-controller-857f4d67dd-4sw9z\" (UID: \"080d2cc3-9d0f-4c08-a279-ab67b3877635\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.929432 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c33e89c0-7bb3-4e34-a770-265dfc317628-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-c7q29\" (UID: \"c33e89c0-7bb3-4e34-a770-265dfc317628\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.934456 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flfs7\" (UniqueName: \"kubernetes.io/projected/83a1bf03-d0e1-4608-9f20-080bf54a2dbc-kube-api-access-flfs7\") pod \"openshift-controller-manager-operator-756b6f6bc6-sw779\" (UID: \"83a1bf03-d0e1-4608-9f20-080bf54a2dbc\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.936848 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p82jv\" (UniqueName: \"kubernetes.io/projected/dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f-kube-api-access-p82jv\") pod \"olm-operator-6b444d44fb-sfxdt\" (UID: \"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.937599 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df1cb7b9-01cf-4b52-9556-3f84680dd057-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6wxts\" (UID: \"df1cb7b9-01cf-4b52-9556-3f84680dd057\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.939007 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.940156 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74kv4\" (UniqueName: \"kubernetes.io/projected/e6bfa0a4-9383-49fe-8fba-3f2bef90dc31-kube-api-access-74kv4\") pod \"dns-default-kw6ql\" (UID: \"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31\") " pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.947601 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnck8\" (UniqueName: \"kubernetes.io/projected/7f36c145-b6cf-42e1-be98-11521fa8d1f9-kube-api-access-jnck8\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4s9c\" (UID: \"7f36c145-b6cf-42e1-be98-11521fa8d1f9\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.948124 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9c5q\" (UniqueName: \"kubernetes.io/projected/64578484-1e9f-49cd-808b-8f036bb9bd48-kube-api-access-b9c5q\") pod \"service-ca-operator-777779d784-cx8vw\" (UID: \"64578484-1e9f-49cd-808b-8f036bb9bd48\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.949755 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nfsp\" (UniqueName: \"kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp\") pod \"collect-profiles-29456850-jlf87\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.954287 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7bbd\" (UniqueName: \"kubernetes.io/projected/50b7d4a0-868d-4ce3-a193-974ad0ef8837-kube-api-access-l7bbd\") pod \"catalog-operator-68c6474976-frwd8\" (UID: \"50b7d4a0-868d-4ce3-a193-974ad0ef8837\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.954347 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:23 crc kubenswrapper[4921]: E0103 03:43:23.954863 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.454386182 +0000 UTC m=+140.065813006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.955076 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.955567 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2wd4\" (UniqueName: \"kubernetes.io/projected/a251d145-2afd-42ec-9a96-4f4fc87a5a1f-kube-api-access-x2wd4\") pod \"machine-config-server-sbgkw\" (UID: \"a251d145-2afd-42ec-9a96-4f4fc87a5a1f\") " pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:23 crc kubenswrapper[4921]: E0103 03:43:23.955762 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.455749239 +0000 UTC m=+140.067176063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.976634 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58lg2\" (UniqueName: \"kubernetes.io/projected/59f35ae7-e556-4393-bb87-eb9601eb36bb-kube-api-access-58lg2\") pod \"machine-config-operator-74547568cd-fts6n\" (UID: \"59f35ae7-e556-4393-bb87-eb9601eb36bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:23 crc kubenswrapper[4921]: I0103 03:43:23.986573 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv6ft\" (UniqueName: \"kubernetes.io/projected/8fa84952-f5e1-4a40-b263-31b8d6a8e361-kube-api-access-fv6ft\") pod \"csi-hostpathplugin-k4ffk\" (UID: \"8fa84952-f5e1-4a40-b263-31b8d6a8e361\") " pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.000582 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.006686 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.009698 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.015583 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/15f2cfb6-4c1b-4328-af51-c0cd466aef3f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-g7wvw\" (UID: \"15f2cfb6-4c1b-4328-af51-c0cd466aef3f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.043533 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.056361 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.056688 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.55665447 +0000 UTC m=+140.168081294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.059436 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.069126 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.076997 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9tl8f" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.097789 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.104945 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.113452 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-sbgkw" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.114912 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.131067 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.163061 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.192572 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.193677 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.197842 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.697787598 +0000 UTC m=+140.309214432 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.199612 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.211016 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.214190 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.222218 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.230181 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.298841 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.299038 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.798999768 +0000 UTC m=+140.410426592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.299467 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.299925 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.799902353 +0000 UTC m=+140.411329177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.319315 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.387289 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.393148 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lsjpr"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.402901 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.403171 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.903123438 +0000 UTC m=+140.514550262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.403448 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.403884 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:24.903866069 +0000 UTC m=+140.515292893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.418949 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.461784 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qr9zc"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.504698 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.505635 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.005611563 +0000 UTC m=+140.617038377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: W0103 03:43:24.529041 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf618317d_4a5b_465b_b14c_72ae7ed50c9c.slice/crio-a12321e6ba9b719ba5169636152de1f68706cd944babc2d1ff786ca05db5d78e WatchSource:0}: Error finding container a12321e6ba9b719ba5169636152de1f68706cd944babc2d1ff786ca05db5d78e: Status 404 returned error can't find the container with id a12321e6ba9b719ba5169636152de1f68706cd944babc2d1ff786ca05db5d78e Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.607930 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.608365 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.108348515 +0000 UTC m=+140.719775339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.660858 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.671720 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2855s"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.676974 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.710676 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.711575 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.21154127 +0000 UTC m=+140.822968094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.717851 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.718360 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.218341378 +0000 UTC m=+140.829768202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: W0103 03:43:24.781119 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5697e250_01f6_4047_8399_ca18e3c17cd5.slice/crio-03368691e311a6023efb3230af6abfe474a7e156bf5c37e28b8b07322c0d11a9 WatchSource:0}: Error finding container 03368691e311a6023efb3230af6abfe474a7e156bf5c37e28b8b07322c0d11a9: Status 404 returned error can't find the container with id 03368691e311a6023efb3230af6abfe474a7e156bf5c37e28b8b07322c0d11a9 Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.786407 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9j876"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.800044 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.820604 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.820644 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.821222 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.821723 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.321703567 +0000 UTC m=+140.933130391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.828250 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.867700 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" event={"ID":"8c06a2f0-3608-4b31-b52b-db33781f3dfa","Type":"ContainerStarted","Data":"8b1646238354ff8601733efe5ffc25147c15d1fae9edbec8c4b4b3218d4b45b6"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.869191 4921 generic.go:334] "Generic (PLEG): container finished" podID="54a51409-4f6d-4c24-9242-a4274fe755a7" containerID="78a220efb7377f15219b53937c6440a062f806240acfe833a5f521ef3b6ce2ca" exitCode=0 Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.869248 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" event={"ID":"54a51409-4f6d-4c24-9242-a4274fe755a7","Type":"ContainerDied","Data":"78a220efb7377f15219b53937c6440a062f806240acfe833a5f521ef3b6ce2ca"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.870859 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" event={"ID":"4c87f43b-b9e0-4c92-985e-3dbad79ff859","Type":"ContainerStarted","Data":"b0d67aa5d43d8f9f5d8a5d1c9c3a804873dc26d84cd55a2f8f0665e8e3d4f015"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.871774 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:24 crc kubenswrapper[4921]: W0103 03:43:24.918057 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod393f5b19_1c3d_4521_8cda_002f27f95734.slice/crio-dc511cc97fae3e4eacaaac4e1f8a5a296b037b01eba51df401fafa47d424febe WatchSource:0}: Error finding container dc511cc97fae3e4eacaaac4e1f8a5a296b037b01eba51df401fafa47d424febe: Status 404 returned error can't find the container with id dc511cc97fae3e4eacaaac4e1f8a5a296b037b01eba51df401fafa47d424febe Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.919211 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" event={"ID":"51c2ba01-1111-4f93-af2f-a6d37f756ace","Type":"ContainerStarted","Data":"7141d87170ec9f5985a57f5e0fbbf8ee784278d58c3731be543d35076395b7df"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.919909 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" event={"ID":"935ca01a-42e9-4cfc-9084-c1ded2c63949","Type":"ContainerStarted","Data":"66e60183ebc4668124806101a71bc5ab328391ef918e00b3f47aac277a88ad0c"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.925063 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:24 crc kubenswrapper[4921]: E0103 03:43:24.925595 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.425579581 +0000 UTC m=+141.037006405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.931553 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" event={"ID":"a2245ae0-ac63-4503-bbd5-59841ca0e9a6","Type":"ContainerStarted","Data":"c89a36e9700838361388fbfe560ab5005bf7b4c23b4493348c5b38ef0ade8810"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.936771 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87"] Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.984954 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" event={"ID":"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0","Type":"ContainerStarted","Data":"b81699d01c48b66596ffe1d5b3b0ede912fbc19d4a940e7210195198fac7efdd"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.987492 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" event={"ID":"f618317d-4a5b-465b-b14c-72ae7ed50c9c","Type":"ContainerStarted","Data":"a12321e6ba9b719ba5169636152de1f68706cd944babc2d1ff786ca05db5d78e"} Jan 03 03:43:24 crc kubenswrapper[4921]: I0103 03:43:24.991391 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" event={"ID":"5697e250-01f6-4047-8399-ca18e3c17cd5","Type":"ContainerStarted","Data":"03368691e311a6023efb3230af6abfe474a7e156bf5c37e28b8b07322c0d11a9"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.020231 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-gqd64" podStartSLOduration=122.020208778 podStartE2EDuration="2m2.020208778s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.016472595 +0000 UTC m=+140.627899429" watchObservedRunningTime="2026-01-03 03:43:25.020208778 +0000 UTC m=+140.631635602" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.041980 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.042207 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.542184798 +0000 UTC m=+141.153611622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.043965 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.059391 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.559366615 +0000 UTC m=+141.170793439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.062494 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x8wkt" event={"ID":"13f1ddbe-b27a-4db0-bd05-a37b298bdebd","Type":"ContainerStarted","Data":"dd8dd947319db8df7c2db3225e45f59a481329dc33af411f5f9e232d54baf19a"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.078289 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.124215 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" event={"ID":"478cd3d6-1d5e-4fe0-a103-2caa602ed743","Type":"ContainerStarted","Data":"06dfe0b5e830b32ec9c4fbcc113827df113685346c6c489c0077eef50c20273f"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.134901 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" event={"ID":"a2fc9836-7783-4f2d-82e3-4750d96b5f07","Type":"ContainerStarted","Data":"de3d9656f951c8e6a7499d01edef3a4184853047833823b239ac73b2d619b1fb"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.148860 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.149354 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.649333322 +0000 UTC m=+141.260760136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.161365 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" event={"ID":"2336e873-d98d-4dc4-81f0-9327afe9d200","Type":"ContainerStarted","Data":"61fea0f1a001e7318d80ad348bf0ce0b3605cadb2925ee7c5bd32c982c6bf5b5"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.172211 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" event={"ID":"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe","Type":"ContainerStarted","Data":"a7b8016a70234d8ccb407034822bef80b48ff44eaca51a33e2d34d10cf02a42e"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.193230 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-62z7t" event={"ID":"c4b61ff0-15aa-4a89-8bf3-e4d6dda31770","Type":"ContainerStarted","Data":"248e7ce701ff555d70874dcce7ce6f7d662fe261884cc54a3d23e5b27afb1bc0"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.194508 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.208677 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" event={"ID":"1d74e762-4edc-473e-b643-af99acf8f210","Type":"ContainerStarted","Data":"6e4128e707bd486ad8e6b8d24afbe639e38f3b5bb4084f05a960af43925113c5"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.209237 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6psdt" podStartSLOduration=121.209211595 podStartE2EDuration="2m1.209211595s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.201350027 +0000 UTC m=+140.812776861" watchObservedRunningTime="2026-01-03 03:43:25.209211595 +0000 UTC m=+140.820638419" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.210768 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.217244 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sbgkw" event={"ID":"a251d145-2afd-42ec-9a96-4f4fc87a5a1f","Type":"ContainerStarted","Data":"3e3ee1868a467b049f4d69d06b66eca11d3389476f66bc41406db95190b9919c"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.218872 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9tl8f"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.235971 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" event={"ID":"66b61c47-d1cd-4c4c-b335-8e20bfd70f5a","Type":"ContainerStarted","Data":"eb4580119d28d410bc8288c3a51da1b2cdf1f573afc8ce3d35bdcb64c7d72f0f"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.244160 4921 generic.go:334] "Generic (PLEG): container finished" podID="ab4a6d77-7628-46ae-9304-da7a2149c336" containerID="17d1c797e85b53f788c6bb6ec82d969272c1b8b5c39bde4a2df8c2e1a69235d9" exitCode=0 Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.245619 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" event={"ID":"ab4a6d77-7628-46ae-9304-da7a2149c336","Type":"ContainerDied","Data":"17d1c797e85b53f788c6bb6ec82d969272c1b8b5c39bde4a2df8c2e1a69235d9"} Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.245770 4921 patch_prober.go:28] interesting pod/downloads-7954f5f757-gqd64 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.245810 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gqd64" podUID="9fa8138b-e464-4c9e-8db6-09e0f03e8507" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.249162 4921 patch_prober.go:28] interesting pod/console-operator-58897d9998-62z7t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.249288 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-62z7t" podUID="c4b61ff0-15aa-4a89-8bf3-e4d6dda31770" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.251506 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.251913 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.75189875 +0000 UTC m=+141.363325574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.261118 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.264863 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-2dn4r" podStartSLOduration=121.264837648 podStartE2EDuration="2m1.264837648s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.264032546 +0000 UTC m=+140.875459370" watchObservedRunningTime="2026-01-03 03:43:25.264837648 +0000 UTC m=+140.876264472" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.352665 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.356754 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.856719079 +0000 UTC m=+141.468146073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.361958 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.367611 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.86756684 +0000 UTC m=+141.478993664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.441047 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" podStartSLOduration=121.441018839 podStartE2EDuration="2m1.441018839s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.29802077 +0000 UTC m=+140.909447604" watchObservedRunningTime="2026-01-03 03:43:25.441018839 +0000 UTC m=+141.052445663" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.465230 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.475906 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:25.975881257 +0000 UTC m=+141.587308081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.499847 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.529907 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.577735 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.578206 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.078186367 +0000 UTC m=+141.689613201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.627670 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-62z7t" podStartSLOduration=122.62765133 podStartE2EDuration="2m2.62765133s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.623846515 +0000 UTC m=+141.235273339" watchObservedRunningTime="2026-01-03 03:43:25.62765133 +0000 UTC m=+141.239078154" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.679038 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.679245 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.179223812 +0000 UTC m=+141.790650626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.679708 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.692724 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.192666085 +0000 UTC m=+141.804092909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.783253 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.789193 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.289078812 +0000 UTC m=+141.900505636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.831405 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4sw9z"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.857481 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-k4ffk"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.872130 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4hxgb" podStartSLOduration=122.872109996 podStartE2EDuration="2m2.872109996s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.86647984 +0000 UTC m=+141.477906664" watchObservedRunningTime="2026-01-03 03:43:25.872109996 +0000 UTC m=+141.483536820" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.891914 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.892548 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.392530413 +0000 UTC m=+142.003957237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.910232 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qmx7w" podStartSLOduration=122.910210364 podStartE2EDuration="2m2.910210364s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:25.907987422 +0000 UTC m=+141.519414246" watchObservedRunningTime="2026-01-03 03:43:25.910210364 +0000 UTC m=+141.521637188" Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.941581 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.945897 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc"] Jan 03 03:43:25 crc kubenswrapper[4921]: I0103 03:43:25.992946 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:25 crc kubenswrapper[4921]: E0103 03:43:25.993399 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.493378132 +0000 UTC m=+142.104804956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.020208 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts"] Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.020306 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-kw6ql"] Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.071893 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29"] Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.073284 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" podStartSLOduration=122.07325871 podStartE2EDuration="2m2.07325871s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.060199107 +0000 UTC m=+141.671625931" watchObservedRunningTime="2026-01-03 03:43:26.07325871 +0000 UTC m=+141.684685534" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.095448 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.096414 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.596390732 +0000 UTC m=+142.207817556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.140608 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw"] Jan 03 03:43:26 crc kubenswrapper[4921]: W0103 03:43:26.178447 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6bfa0a4_9383_49fe_8fba_3f2bef90dc31.slice/crio-f7dfa4d27c9d6c96a9f59d6e7d90ebfb5da02a0994d7e1f2662b6a6e31210d13 WatchSource:0}: Error finding container f7dfa4d27c9d6c96a9f59d6e7d90ebfb5da02a0994d7e1f2662b6a6e31210d13: Status 404 returned error can't find the container with id f7dfa4d27c9d6c96a9f59d6e7d90ebfb5da02a0994d7e1f2662b6a6e31210d13 Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.192243 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c"] Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.197256 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.198102 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.698055705 +0000 UTC m=+142.309482529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.198461 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.199413 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.699405172 +0000 UTC m=+142.310831996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.238020 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x"] Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.300862 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.301555 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.801492265 +0000 UTC m=+142.412919089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.302609 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.303674 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.803656925 +0000 UTC m=+142.415083749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.321032 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" event={"ID":"83a1bf03-d0e1-4608-9f20-080bf54a2dbc","Type":"ContainerStarted","Data":"13d6af05bf33880a48279597bcca61b8975b421af913203f6b35119f4c2e309a"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.325614 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" event={"ID":"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f","Type":"ContainerStarted","Data":"fe4dc2a91e3f73a4b015eb2082f84ca5890f8d544f9e5e4e7b0a334595fdf285"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.390241 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" event={"ID":"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe","Type":"ContainerStarted","Data":"7c2d7fb80d8645c72ea1e02509c2d8af4811ec19408bfa5849277e1f413d7597"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.409942 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.410449 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:26.910418369 +0000 UTC m=+142.521845213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.441889 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" event={"ID":"f618317d-4a5b-465b-b14c-72ae7ed50c9c","Type":"ContainerStarted","Data":"9251f1e7ca7e819fa1e3775444ac1187c809401e6123d2dc0222685aa0253938"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.448223 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.449652 4921 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-m7q7x container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.449733 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" podUID="f618317d-4a5b-465b-b14c-72ae7ed50c9c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.525717 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.529685 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.029657589 +0000 UTC m=+142.641084413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.554133 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x8wkt" event={"ID":"13f1ddbe-b27a-4db0-bd05-a37b298bdebd","Type":"ContainerStarted","Data":"452c33f15967d9ac985c3a871df43e73c36cba8fec7846a7e531ddb7e6921c19"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.565058 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" podStartSLOduration=122.56502119 podStartE2EDuration="2m2.56502119s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.496326294 +0000 UTC m=+142.107753118" watchObservedRunningTime="2026-01-03 03:43:26.56502119 +0000 UTC m=+142.176448004" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.607741 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" event={"ID":"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c","Type":"ContainerStarted","Data":"2b8ed6e0ac576916736107a464cfc0e665365cac1d4f8857774a1c313f4ffb49"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.633153 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.633210 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-x8wkt" podStartSLOduration=122.633183893 podStartE2EDuration="2m2.633183893s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.630810777 +0000 UTC m=+142.242237601" watchObservedRunningTime="2026-01-03 03:43:26.633183893 +0000 UTC m=+142.244610717" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.634387 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.134367146 +0000 UTC m=+142.745793970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.658665 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" event={"ID":"64578484-1e9f-49cd-808b-8f036bb9bd48","Type":"ContainerStarted","Data":"446431205d791084ad60699bcd0a83574ca021896acb52bc61f4d9cd4b9d4713"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.658997 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" event={"ID":"64578484-1e9f-49cd-808b-8f036bb9bd48","Type":"ContainerStarted","Data":"2f210864695021ffe6bff491978aaac300729855f37107049941312f1d99cb97"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.729955 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cx8vw" podStartSLOduration=122.729934978 podStartE2EDuration="2m2.729934978s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.727417898 +0000 UTC m=+142.338844722" watchObservedRunningTime="2026-01-03 03:43:26.729934978 +0000 UTC m=+142.341361802" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.734734 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.735082 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.235067541 +0000 UTC m=+142.846494365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.755187 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" event={"ID":"5697e250-01f6-4047-8399-ca18e3c17cd5","Type":"ContainerStarted","Data":"7ff3a7ff93854cd2e89783f1e14e35aa2672291c0490b6e0f89167c2bf1dab6a"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.780255 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" event={"ID":"8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0","Type":"ContainerStarted","Data":"5bf185a9de5c21626194210709584e9b8401e6f6b854870d19b10400403551e4"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.805699 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-sbgkw" event={"ID":"a251d145-2afd-42ec-9a96-4f4fc87a5a1f","Type":"ContainerStarted","Data":"d4bb3d8b4cc1e26b63cf8f02c10bf2ce7d5bf18f3fdeef786ccc3d7009b4939f"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.829514 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-fxkcl" podStartSLOduration=122.829494122 podStartE2EDuration="2m2.829494122s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.821760348 +0000 UTC m=+142.433187172" watchObservedRunningTime="2026-01-03 03:43:26.829494122 +0000 UTC m=+142.440920946" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.833697 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" event={"ID":"a2245ae0-ac63-4503-bbd5-59841ca0e9a6","Type":"ContainerStarted","Data":"56e6a0e26a3b304de6737555ee48197e51a3424a21633adcfb6410eaccf79d94"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.841124 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.842205 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.342186814 +0000 UTC m=+142.953613638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.859706 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" event={"ID":"714b3568-8ca6-4d1a-b451-1dae7b24fbdc","Type":"ContainerStarted","Data":"fcd55755570d4ab6ac3b134ecc4b07a93d9869ea7d9bd35cb0925a98e6789a8d"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.861849 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9j876" event={"ID":"393f5b19-1c3d-4521-8cda-002f27f95734","Type":"ContainerStarted","Data":"053ea92955e37141d8080b68fb46ecb8991afc4eb0bd73b3df2c1615a49fa69d"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.861872 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9j876" event={"ID":"393f5b19-1c3d-4521-8cda-002f27f95734","Type":"ContainerStarted","Data":"dc511cc97fae3e4eacaaac4e1f8a5a296b037b01eba51df401fafa47d424febe"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.917428 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" event={"ID":"080d2cc3-9d0f-4c08-a279-ab67b3877635","Type":"ContainerStarted","Data":"834a383a46d704a55b8567ec9813eebecbbf74834b44a5dd8e4d281530a35513"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.943473 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:26 crc kubenswrapper[4921]: E0103 03:43:26.944884 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.444868615 +0000 UTC m=+143.056295439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.957041 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-sbgkw" podStartSLOduration=6.957008262 podStartE2EDuration="6.957008262s" podCreationTimestamp="2026-01-03 03:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.924692625 +0000 UTC m=+142.536119449" watchObservedRunningTime="2026-01-03 03:43:26.957008262 +0000 UTC m=+142.568435086" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.968594 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" event={"ID":"50b7d4a0-868d-4ce3-a193-974ad0ef8837","Type":"ContainerStarted","Data":"c19e805ddb9a07b291be153698476ae82c440b87d50faab8c14751df4e54bbde"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.969786 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.971198 4921 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-frwd8 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.971263 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" podUID="50b7d4a0-868d-4ce3-a193-974ad0ef8837" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.976014 4921 generic.go:334] "Generic (PLEG): container finished" podID="a2fc9836-7783-4f2d-82e3-4750d96b5f07" containerID="6b48db8c8a994508eb311321276529e2c021b0597d11d8770fd23e8826a80794" exitCode=0 Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.976080 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" event={"ID":"a2fc9836-7783-4f2d-82e3-4750d96b5f07","Type":"ContainerDied","Data":"6b48db8c8a994508eb311321276529e2c021b0597d11d8770fd23e8826a80794"} Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.984992 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.994600 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:26 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:26 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:26 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:26 crc kubenswrapper[4921]: I0103 03:43:26.994951 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.006205 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9j876" podStartSLOduration=123.006179907 podStartE2EDuration="2m3.006179907s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:26.986713987 +0000 UTC m=+142.598140811" watchObservedRunningTime="2026-01-03 03:43:27.006179907 +0000 UTC m=+142.617606731" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.045204 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.046826 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.546801144 +0000 UTC m=+143.158227968 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.071762 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" event={"ID":"8c06a2f0-3608-4b31-b52b-db33781f3dfa","Type":"ContainerStarted","Data":"e4b3743478d8831b4d8672f7e93d858e303d848e0f894b87dc8f48e5cfc04bf8"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.073078 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.082981 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" podStartSLOduration=123.082950098 podStartE2EDuration="2m3.082950098s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:27.022061188 +0000 UTC m=+142.633488012" watchObservedRunningTime="2026-01-03 03:43:27.082950098 +0000 UTC m=+142.694376922" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.083593 4921 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zng4m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.083680 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.153020 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.154971 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.654955947 +0000 UTC m=+143.266382771 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.171872 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" event={"ID":"16503625-6d0f-42a8-b7fb-991ae13be2d4","Type":"ContainerStarted","Data":"e85f2a92a35d47ed96da2cfd2411ab02868826f544151624267b5a1156f2f1b0"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.178867 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" podStartSLOduration=123.17884191 podStartE2EDuration="2m3.17884191s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:27.171764644 +0000 UTC m=+142.783191468" watchObservedRunningTime="2026-01-03 03:43:27.17884191 +0000 UTC m=+142.790268734" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.216077 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2855s" event={"ID":"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9","Type":"ContainerStarted","Data":"f92897e73e22679dd1cb03e656d0be7664cc969961c02eefb91ad5441536ef41"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.216135 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2855s" event={"ID":"2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9","Type":"ContainerStarted","Data":"1f2f254e9fbc9aa61962eb8f6d45953841a854d8aaddffe6aff6041f624996f2"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.241810 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" event={"ID":"8f625709-e003-4472-990b-7d7bb1c81477","Type":"ContainerStarted","Data":"09dec84db6c2629d7610eb6c5491250427ea0f17a160dc8ec4252f7a7bb0164a"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.256262 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.258216 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.758178602 +0000 UTC m=+143.369605426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.258498 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9tl8f" event={"ID":"c2e8f120-d9bf-4f9b-960e-5f816329d194","Type":"ContainerStarted","Data":"72711e85644ad6b2da31ab341b83a78fcc31a943b295b7a0aabd34dc1175d8a5"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.299220 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" event={"ID":"8fa84952-f5e1-4a40-b263-31b8d6a8e361","Type":"ContainerStarted","Data":"7194fb8f9a4892e496a4e59e43552c1edd9ec97d728f99cdaf8bbed346fdcda4"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.299680 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2855s" podStartSLOduration=124.299655204 podStartE2EDuration="2m4.299655204s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:27.298862431 +0000 UTC m=+142.910289255" watchObservedRunningTime="2026-01-03 03:43:27.299655204 +0000 UTC m=+142.911082028" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.337524 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" event={"ID":"59f35ae7-e556-4393-bb87-eb9601eb36bb","Type":"ContainerStarted","Data":"717a736d2b8a1aa0df27ca617456020b3ebbea4030c8b99768a05a7de7927b8e"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.350556 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-9tl8f" podStartSLOduration=7.350537716 podStartE2EDuration="7.350537716s" podCreationTimestamp="2026-01-03 03:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:27.349876598 +0000 UTC m=+142.961303422" watchObservedRunningTime="2026-01-03 03:43:27.350537716 +0000 UTC m=+142.961964540" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.358225 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.358962 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.858913249 +0000 UTC m=+143.470340103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.392551 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" event={"ID":"dc0605d6-fd4e-41cd-b8f6-27a304c252c0","Type":"ContainerStarted","Data":"7247476ecae5b04b6253c5c591460f3e3ca3ebe15856473f3dd123466b9dea64"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.397298 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6ql" event={"ID":"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31","Type":"ContainerStarted","Data":"f7dfa4d27c9d6c96a9f59d6e7d90ebfb5da02a0994d7e1f2662b6a6e31210d13"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.404839 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" event={"ID":"df1cb7b9-01cf-4b52-9556-3f84680dd057","Type":"ContainerStarted","Data":"09d03eeaa0aef926b17ada2f3ecc97333ef4e6fa8226cdd945f2fe017d578623"} Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.407308 4921 patch_prober.go:28] interesting pod/downloads-7954f5f757-gqd64 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.407389 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gqd64" podUID="9fa8138b-e464-4c9e-8db6-09e0f03e8507" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.415240 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-62z7t" Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.460837 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.463524 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:27.963490571 +0000 UTC m=+143.574917395 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.564805 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.565759 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.06574244 +0000 UTC m=+143.677169264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.667377 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.667862 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.167838304 +0000 UTC m=+143.779265128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.669886 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.670434 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.170422506 +0000 UTC m=+143.781849330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.772059 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.772484 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.272462579 +0000 UTC m=+143.883889403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.873885 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.875205 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.37518772 +0000 UTC m=+143.986614544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:27 crc kubenswrapper[4921]: I0103 03:43:27.975537 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:27 crc kubenswrapper[4921]: E0103 03:43:27.975928 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.475911386 +0000 UTC m=+144.087338210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.005088 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:28 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:28 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:28 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.005169 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.077207 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.078068 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.578049612 +0000 UTC m=+144.189476436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.178587 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.178998 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.678977163 +0000 UTC m=+144.290403987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.282707 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.283070 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.783054532 +0000 UTC m=+144.394481356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.383873 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.385347 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.885321931 +0000 UTC m=+144.496748765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.476216 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" event={"ID":"080d2cc3-9d0f-4c08-a279-ab67b3877635","Type":"ContainerStarted","Data":"dac2462bb1208c0ae71964f8a9cedcd9abdf6e2b772f43fef2af5c3b4c8127ce"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.485593 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.485938 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:28.985924583 +0000 UTC m=+144.597351407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.497573 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" event={"ID":"15f2cfb6-4c1b-4328-af51-c0cd466aef3f","Type":"ContainerStarted","Data":"ba9eb5c3b01ce1a1b61f5ea150275bc11cc08ab7eef8cb4898ff6a8d17c8f0d2"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.497630 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" event={"ID":"15f2cfb6-4c1b-4328-af51-c0cd466aef3f","Type":"ContainerStarted","Data":"cfc2f1752592251cb8945bcf194312a4f54c597902024a1b01bfbd954d0ba72b"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.514675 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" event={"ID":"935ca01a-42e9-4cfc-9084-c1ded2c63949","Type":"ContainerStarted","Data":"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.515346 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.524431 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.533906 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" event={"ID":"59f35ae7-e556-4393-bb87-eb9601eb36bb","Type":"ContainerStarted","Data":"4fb4f828e2a3e1965e369bbc56468de194c5e85d0cf02c1db6cdd5edf8938d69"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.533964 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" event={"ID":"59f35ae7-e556-4393-bb87-eb9601eb36bb","Type":"ContainerStarted","Data":"4e86e8b18aa26d3d842d553b6149f813d2f1ffda9b5757df387c67e99e179d4a"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.534020 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-g7wvw" podStartSLOduration=124.534008618 podStartE2EDuration="2m4.534008618s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.530207273 +0000 UTC m=+144.141634097" watchObservedRunningTime="2026-01-03 03:43:28.534008618 +0000 UTC m=+144.145435442" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.556540 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" event={"ID":"c33e89c0-7bb3-4e34-a770-265dfc317628","Type":"ContainerStarted","Data":"bce60f752e4df355ee2d6d89ca27d7bd1d9e1085ae41accc75c597ab297297e6"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.556584 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" event={"ID":"c33e89c0-7bb3-4e34-a770-265dfc317628","Type":"ContainerStarted","Data":"bf512a59e8bf53c4f1c35cd0eb950241df2d4433c996a06093fe4046b1887874"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.570399 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" event={"ID":"df1cb7b9-01cf-4b52-9556-3f84680dd057","Type":"ContainerStarted","Data":"75f0dbfc073ecc43156fdd5a74bd08ff8de06055447594319fe29e7842f2854c"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.588558 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-c7q29" podStartSLOduration=124.588543732 podStartE2EDuration="2m4.588543732s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.584359636 +0000 UTC m=+144.195786460" watchObservedRunningTime="2026-01-03 03:43:28.588543732 +0000 UTC m=+144.199970556" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.588716 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.588956 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" podStartSLOduration=125.588952084 podStartE2EDuration="2m5.588952084s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.557518251 +0000 UTC m=+144.168945075" watchObservedRunningTime="2026-01-03 03:43:28.588952084 +0000 UTC m=+144.200378908" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.588979 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.088939113 +0000 UTC m=+144.700365937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.589811 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.593889 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.093874681 +0000 UTC m=+144.705301505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.619998 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" event={"ID":"a2245ae0-ac63-4503-bbd5-59841ca0e9a6","Type":"ContainerStarted","Data":"7329542485e0d7321ab823052e17c045816ce5f819b9cca14b222ccb0d1392f6"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.649323 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" event={"ID":"ab4a6d77-7628-46ae-9304-da7a2149c336","Type":"ContainerStarted","Data":"02b166860ff6e445f1840ed1741b7fd6a8a71d835cd4fa2a8785f1f9d58af69c"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.673607 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" event={"ID":"714b3568-8ca6-4d1a-b451-1dae7b24fbdc","Type":"ContainerStarted","Data":"67a3d0a4671501f907a855f258072a0ab155fbeecad5aad823defe6bddd169e8"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.673659 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" event={"ID":"714b3568-8ca6-4d1a-b451-1dae7b24fbdc","Type":"ContainerStarted","Data":"8d514bd973300e6daea46ef7642129f4ecec204b19c49d022ccd3376e7fc07fb"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.674348 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fts6n" podStartSLOduration=124.674330783 podStartE2EDuration="2m4.674330783s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.672000779 +0000 UTC m=+144.283427603" watchObservedRunningTime="2026-01-03 03:43:28.674330783 +0000 UTC m=+144.285757607" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.691002 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.692341 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.192317663 +0000 UTC m=+144.803744487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.731261 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" event={"ID":"7f36c145-b6cf-42e1-be98-11521fa8d1f9","Type":"ContainerStarted","Data":"449578ac6520a7e04332fe2deae5d10766c8c396856e1970328089b6d8d31921"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.731601 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" event={"ID":"7f36c145-b6cf-42e1-be98-11521fa8d1f9","Type":"ContainerStarted","Data":"d3c22442f6bbd70c7673aca934b50b02edfe8a597a71a0befe831452e2c7cfb3"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.743259 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8wjxs" podStartSLOduration=125.743234986 podStartE2EDuration="2m5.743234986s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.734299649 +0000 UTC m=+144.345726473" watchObservedRunningTime="2026-01-03 03:43:28.743234986 +0000 UTC m=+144.354661810" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.788482 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.789520 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.792859 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.793302 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.794749 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.294733966 +0000 UTC m=+144.906160790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.795899 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" event={"ID":"a2fc9836-7783-4f2d-82e3-4750d96b5f07","Type":"ContainerStarted","Data":"f721b2b8467125682ebdb97f745e73ba2a8ea58ba0a5cb00bf3a399186952479"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.797969 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.825901 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-96kc8" podStartSLOduration=124.825875891 podStartE2EDuration="2m4.825875891s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.810873844 +0000 UTC m=+144.422300668" watchObservedRunningTime="2026-01-03 03:43:28.825875891 +0000 UTC m=+144.437302715" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.826675 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.830807 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" event={"ID":"dffb0a13-d7a3-4ae4-a5d4-1b7f66666f1f","Type":"ContainerStarted","Data":"b88010f43034350a26c8dff0bb1849c5aa84e495fe18aa1009d3186e369f20cf"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.837287 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.855055 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4s9c" podStartSLOduration=124.85502476 podStartE2EDuration="2m4.85502476s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.839012145 +0000 UTC m=+144.450438969" watchObservedRunningTime="2026-01-03 03:43:28.85502476 +0000 UTC m=+144.466451584" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.895213 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.895678 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgjzr\" (UniqueName: \"kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.895725 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.895848 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:28 crc kubenswrapper[4921]: E0103 03:43:28.898745 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.398706893 +0000 UTC m=+145.010133717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.963437 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" event={"ID":"8f625709-e003-4472-990b-7d7bb1c81477","Type":"ContainerStarted","Data":"21e1fcfc1852db7b39e36a6f1c80ecacbd5d5a724a6043832f9c56edde6eb453"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.963521 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.979221 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" podStartSLOduration=125.979198866 podStartE2EDuration="2m5.979198866s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:28.969185399 +0000 UTC m=+144.580612223" watchObservedRunningTime="2026-01-03 03:43:28.979198866 +0000 UTC m=+144.590625690" Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.984258 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" event={"ID":"ed2a16e8-45bd-4bbb-bd03-759dd8c30ffe","Type":"ContainerStarted","Data":"6801eaff5b78a9b655e58fee280b9a4158fcacd6234efdd242367afd3e84ccff"} Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.994039 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:28 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:28 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:28 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:28 crc kubenswrapper[4921]: I0103 03:43:28.994111 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:28.998926 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:28.999363 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:28.999701 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgjzr\" (UniqueName: \"kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:28.999751 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.001108 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.501082694 +0000 UTC m=+145.112509688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.008777 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9tl8f" event={"ID":"c2e8f120-d9bf-4f9b-960e-5f816329d194","Type":"ContainerStarted","Data":"8f1012f582e4bc7afd964fca59b2986560ec39f8869e14d1035955054a16def4"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.017582 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.020097 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.033463 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" event={"ID":"dc0605d6-fd4e-41cd-b8f6-27a304c252c0","Type":"ContainerStarted","Data":"e6537df77e9d4f5807046c07a90cf83d797e331d7ccaeec0c32bc09705a43cdb"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.033520 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" event={"ID":"dc0605d6-fd4e-41cd-b8f6-27a304c252c0","Type":"ContainerStarted","Data":"136009917d6c9cda7b3696193be0243d1a544e2f2385697d7c47cf901e1cd0d1"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.043732 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6ql" event={"ID":"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31","Type":"ContainerStarted","Data":"58d4b66a43816a5f375bb85fa781cf98fcb0e29a57cfa844455674d5a3dff944"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.061414 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" event={"ID":"55705621-48c6-4662-b10c-492b16cb1368","Type":"ContainerStarted","Data":"4a9f65864ea781b52fa5dfada6ae87f1d519e7209b8a62af225dd4c4d64601b9"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.079796 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.090989 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" event={"ID":"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c","Type":"ContainerStarted","Data":"b2e39a43b31941ad6b4ac11ce08f580170f306366a0238d925cd36ec0a92f373"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.091192 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.091557 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" event={"ID":"50b7d4a0-868d-4ce3-a193-974ad0ef8837","Type":"ContainerStarted","Data":"623f7986140aa78f2f3348e744cd5ab7a4a341a7c503b959cb40cce4aa5914af"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.104919 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.105961 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.106358 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.106387 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7hpf\" (UniqueName: \"kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.106553 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.106670 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.606650845 +0000 UTC m=+145.218077669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.136711 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" event={"ID":"5697e250-01f6-4047-8399-ca18e3c17cd5","Type":"ContainerStarted","Data":"cb6f8c326f4f93a48eeea7b8c8211c0fcd24585d2875810ed01aeb3e06666847"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.137633 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.139662 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgjzr\" (UniqueName: \"kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr\") pod \"community-operators-qh9w2\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.162973 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-frwd8" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.166040 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" event={"ID":"54a51409-4f6d-4c24-9242-a4274fe755a7","Type":"ContainerStarted","Data":"38617716d9bf194dd0ce230199596495acd8e424687d6e6418042f1a948711c3"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.178857 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.190835 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" event={"ID":"83a1bf03-d0e1-4608-9f20-080bf54a2dbc","Type":"ContainerStarted","Data":"857a5d50527df12e73d4fc7cd75089f5823dda2596093669240dcec20778e4e8"} Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.197255 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.198249 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-sfxdt" podStartSLOduration=125.198231116 podStartE2EDuration="2m5.198231116s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.196087138 +0000 UTC m=+144.807513962" watchObservedRunningTime="2026-01-03 03:43:29.198231116 +0000 UTC m=+144.809657940" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.206315 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.210094 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.210207 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.210226 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7hpf\" (UniqueName: \"kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.210693 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.214361 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.714342084 +0000 UTC m=+145.325768908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.218511 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-m7q7x" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.220957 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.221523 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.313675 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.314081 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.814060972 +0000 UTC m=+145.425487796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.324897 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.327053 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.340448 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7hpf\" (UniqueName: \"kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf\") pod \"certified-operators-jj6h6\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.344342 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" podStartSLOduration=125.344290282 podStartE2EDuration="2m5.344290282s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.339760226 +0000 UTC m=+144.951187050" watchObservedRunningTime="2026-01-03 03:43:29.344290282 +0000 UTC m=+144.955717106" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.351519 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.414980 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.415027 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.415140 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grqcz\" (UniqueName: \"kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.415169 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.415559 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:29.915545899 +0000 UTC m=+145.526972723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.450641 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.461305 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.462469 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519097 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519473 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519556 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grqcz\" (UniqueName: \"kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519587 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519633 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519654 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.519673 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgqq\" (UniqueName: \"kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.519819 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.019799243 +0000 UTC m=+145.631226067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.520503 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.521310 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.521668 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.623537 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.623653 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.623691 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.623737 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgqq\" (UniqueName: \"kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.624793 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.625086 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.625242 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grqcz\" (UniqueName: \"kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz\") pod \"community-operators-cpcqj\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.625463 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.125446446 +0000 UTC m=+145.736873440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.657958 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.661061 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" podStartSLOduration=125.661033734 podStartE2EDuration="2m5.661033734s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.632909554 +0000 UTC m=+145.244336378" watchObservedRunningTime="2026-01-03 03:43:29.661033734 +0000 UTC m=+145.272460558" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.725549 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.726316 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.226299086 +0000 UTC m=+145.837725910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.751794 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgqq\" (UniqueName: \"kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq\") pod \"certified-operators-5lkzl\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.773305 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-lsjpr" podStartSLOduration=125.773249349 podStartE2EDuration="2m5.773249349s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.758244172 +0000 UTC m=+145.369671006" watchObservedRunningTime="2026-01-03 03:43:29.773249349 +0000 UTC m=+145.384676173" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.820684 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.829340 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.830112 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.330073336 +0000 UTC m=+145.941500160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.859573 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" podStartSLOduration=125.859541995 podStartE2EDuration="2m5.859541995s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.819481322 +0000 UTC m=+145.430908146" watchObservedRunningTime="2026-01-03 03:43:29.859541995 +0000 UTC m=+145.470968819" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.907843 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" podStartSLOduration=125.907814585 podStartE2EDuration="2m5.907814585s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.865941572 +0000 UTC m=+145.477368396" watchObservedRunningTime="2026-01-03 03:43:29.907814585 +0000 UTC m=+145.519241409" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.934106 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:29 crc kubenswrapper[4921]: E0103 03:43:29.934563 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.434536646 +0000 UTC m=+146.045963480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.955673 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sw779" podStartSLOduration=125.955651642 podStartE2EDuration="2m5.955651642s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.9537899 +0000 UTC m=+145.565216724" watchObservedRunningTime="2026-01-03 03:43:29.955651642 +0000 UTC m=+145.567078466" Jan 03 03:43:29 crc kubenswrapper[4921]: I0103 03:43:29.956626 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-jvxlw" podStartSLOduration=125.956619049 podStartE2EDuration="2m5.956619049s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:29.911758104 +0000 UTC m=+145.523184928" watchObservedRunningTime="2026-01-03 03:43:29.956619049 +0000 UTC m=+145.568045873" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.007541 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:30 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:30 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:30 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.007627 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.044868 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.045246 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.545230079 +0000 UTC m=+146.156656903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.150471 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.150948 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.650930484 +0000 UTC m=+146.262357308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.239704 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" event={"ID":"df1cb7b9-01cf-4b52-9556-3f84680dd057","Type":"ContainerStarted","Data":"3f8b774ebadf6b3c2b9c1f596a294b75a72f55e7e6c7271c121da258d5c8a462"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.253429 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.255434 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.755413344 +0000 UTC m=+146.366840168 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.287343 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6wxts" podStartSLOduration=126.287317289 podStartE2EDuration="2m6.287317289s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.27403585 +0000 UTC m=+145.885462674" watchObservedRunningTime="2026-01-03 03:43:30.287317289 +0000 UTC m=+145.898744113" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.313098 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" event={"ID":"8f625709-e003-4472-990b-7d7bb1c81477","Type":"ContainerStarted","Data":"d42b6c2eb0fc337253c3321213e0d3cbb07f115a1b0448954ec55438b0c93196"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.347699 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-trzp5" podStartSLOduration=127.347672065 podStartE2EDuration="2m7.347672065s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.342497231 +0000 UTC m=+145.953924045" watchObservedRunningTime="2026-01-03 03:43:30.347672065 +0000 UTC m=+145.959098889" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.356856 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" event={"ID":"ab4a6d77-7628-46ae-9304-da7a2149c336","Type":"ContainerStarted","Data":"7a2a927013860562032574c9537de4a4b60cad2677435030593eb243185dee83"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.358741 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.360067 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.860042828 +0000 UTC m=+146.471469652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.385470 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" event={"ID":"8fa84952-f5e1-4a40-b263-31b8d6a8e361","Type":"ContainerStarted","Data":"f6ef810a9886b4b69c609be37aac1f6b360ebd464689b9968a0e337fecc1c593"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.401665 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" podStartSLOduration=127.401646173 podStartE2EDuration="2m7.401646173s" podCreationTimestamp="2026-01-03 03:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.40011139 +0000 UTC m=+146.011538204" watchObservedRunningTime="2026-01-03 03:43:30.401646173 +0000 UTC m=+146.013072997" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.421746 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.430771 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-zh24x" event={"ID":"55705621-48c6-4662-b10c-492b16cb1368","Type":"ContainerStarted","Data":"e69e724e4d6e23c17d7650cb772ef69c8a674db6b80c68113dd7b94463b61527"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.434293 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" event={"ID":"080d2cc3-9d0f-4c08-a279-ab67b3877635","Type":"ContainerStarted","Data":"36060f6ff7b5f69a43dd440b3b9c00272ab47b37d1290e626dba78d75ea64d16"} Jan 03 03:43:30 crc kubenswrapper[4921]: W0103 03:43:30.438965 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5545afd_0acb_4149_8e56_2804de448a28.slice/crio-85d2bc7e9317e1e4acc8f2cd96bcd1d3953612c03382ef48829d4c58b505acd6 WatchSource:0}: Error finding container 85d2bc7e9317e1e4acc8f2cd96bcd1d3953612c03382ef48829d4c58b505acd6: Status 404 returned error can't find the container with id 85d2bc7e9317e1e4acc8f2cd96bcd1d3953612c03382ef48829d4c58b505acd6 Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.442696 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" event={"ID":"16503625-6d0f-42a8-b7fb-991ae13be2d4","Type":"ContainerStarted","Data":"e5ed7129708d4eb086be38b09dc29c7da34b23c4954658930b5622eb6de9908b"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.461291 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-kw6ql" event={"ID":"e6bfa0a4-9383-49fe-8fba-3f2bef90dc31","Type":"ContainerStarted","Data":"e4b3364c48fb514738d569e15edde7b7bb73f30d25d2fb54b26f59e30aa1b3af"} Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.463724 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.469825 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:30.969803584 +0000 UTC m=+146.581230408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.471217 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.542186 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4sw9z" podStartSLOduration=126.542165153 podStartE2EDuration="2m6.542165153s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.502629866 +0000 UTC m=+146.114056680" watchObservedRunningTime="2026-01-03 03:43:30.542165153 +0000 UTC m=+146.153591977" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.542672 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-kw6ql" podStartSLOduration=10.542668547 podStartE2EDuration="10.542668547s" podCreationTimestamp="2026-01-03 03:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.540623121 +0000 UTC m=+146.152049945" watchObservedRunningTime="2026-01-03 03:43:30.542668547 +0000 UTC m=+146.154095371" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.567783 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.575932 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.075892159 +0000 UTC m=+146.687318983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.600142 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-vb9dc" podStartSLOduration=126.600110992 podStartE2EDuration="2m6.600110992s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:30.589839067 +0000 UTC m=+146.201265891" watchObservedRunningTime="2026-01-03 03:43:30.600110992 +0000 UTC m=+146.211537806" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.624077 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-hsm7c" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.675605 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.677652 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.177637764 +0000 UTC m=+146.789064588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.749643 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.759957 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.762702 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.774170 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.778146 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.796060 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.778257 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.278234746 +0000 UTC m=+146.889661570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.796613 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rtrf\" (UniqueName: \"kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.796805 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.796971 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.798469 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.798984 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.298966162 +0000 UTC m=+146.910392986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.899981 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.900351 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.900449 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rtrf\" (UniqueName: \"kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.900496 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.901010 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: E0103 03:43:30.901120 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.401099877 +0000 UTC m=+147.012526701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.901371 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.961748 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rtrf\" (UniqueName: \"kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf\") pod \"redhat-marketplace-khcdk\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.988819 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:30 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:30 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:30 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:30 crc kubenswrapper[4921]: I0103 03:43:30.988880 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.001658 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.002548 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.502531733 +0000 UTC m=+147.113958557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.019696 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.056103 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.072929 4921 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.102904 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.103071 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.603039142 +0000 UTC m=+147.214465966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.103211 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.103721 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.603704772 +0000 UTC m=+147.215131596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.120712 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.160120 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.163797 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.176568 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.176959 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.179240 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.207830 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.208081 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kqs7\" (UniqueName: \"kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.208165 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.208230 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.208386 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.708362547 +0000 UTC m=+147.319789371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.310921 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.311197 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kqs7\" (UniqueName: \"kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.311236 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.311281 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.311748 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.311973 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.312634 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.81261947 +0000 UTC m=+147.424046294 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.378420 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kqs7\" (UniqueName: \"kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7\") pod \"redhat-marketplace-jwp9b\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.412918 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.413127 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.91309563 +0000 UTC m=+147.524522444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.413394 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.414097 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:31.914081577 +0000 UTC m=+147.525508401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.479821 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerStarted","Data":"d483f8f3737904b6b789efec9b3704d919cabdc6b52c4cac6cb00f6746b637aa"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.479941 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerStarted","Data":"ca2f733d699ff2d8e395df2fd037bb89a0a112d90799b9bdb349b59fe99e1f25"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.484070 4921 generic.go:334] "Generic (PLEG): container finished" podID="a79c2918-942a-4a60-a300-e3981d43dd75" containerID="72f9e0b7124e5eb3cb96a7a9713c397bb31d2df81e7e3a602ecf2e108ccb42bf" exitCode=0 Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.484121 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerDied","Data":"72f9e0b7124e5eb3cb96a7a9713c397bb31d2df81e7e3a602ecf2e108ccb42bf"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.484141 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerStarted","Data":"ce1a842c47a7c8e1e6ee9206c9362969a7cf4d03e515218ec114232ffa14b19a"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.491117 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.501050 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" event={"ID":"8fa84952-f5e1-4a40-b263-31b8d6a8e361","Type":"ContainerStarted","Data":"738ddc6789eebc10c7ba69deb101c8700bac5d830e452e080947f994384631a1"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.501118 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" event={"ID":"8fa84952-f5e1-4a40-b263-31b8d6a8e361","Type":"ContainerStarted","Data":"fb885589bf9333e6196c1b5031d41b90ada6d1e795c78b4682b60584460c3584"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.517573 4921 generic.go:334] "Generic (PLEG): container finished" podID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerID="e534eaa8fa63e3d9bd982c4a5d07280711d3560dff64c0a0ac3b66c164b14533" exitCode=0 Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.517763 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerDied","Data":"e534eaa8fa63e3d9bd982c4a5d07280711d3560dff64c0a0ac3b66c164b14533"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.517823 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerStarted","Data":"762e748c81c0f98162fa75d2c3241047a73b3185fb5fffee0464f9a5e112c4b4"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.518282 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.518606 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-03 03:43:32.018585618 +0000 UTC m=+147.630012442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.533507 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.552613 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.559434 4921 generic.go:334] "Generic (PLEG): container finished" podID="d5545afd-0acb-4149-8e56-2804de448a28" containerID="ed0905f013ed3acfd1db072ee0ca00197fbd85cbbbb08d88de6b53e218a92c9e" exitCode=0 Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.561155 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerDied","Data":"ed0905f013ed3acfd1db072ee0ca00197fbd85cbbbb08d88de6b53e218a92c9e"} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.561204 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerStarted","Data":"85d2bc7e9317e1e4acc8f2cd96bcd1d3953612c03382ef48829d4c58b505acd6"} Jan 03 03:43:31 crc kubenswrapper[4921]: W0103 03:43:31.592629 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb5cde82_e1a0_4d9f_800a_9aa70cc6e728.slice/crio-d110296cbd1a7409b1d96d72ed6446e215987f72708d7e920f1df145f4299dfe WatchSource:0}: Error finding container d110296cbd1a7409b1d96d72ed6446e215987f72708d7e920f1df145f4299dfe: Status 404 returned error can't find the container with id d110296cbd1a7409b1d96d72ed6446e215987f72708d7e920f1df145f4299dfe Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.620799 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: E0103 03:43:31.631399 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-03 03:43:32.131374469 +0000 UTC m=+147.742801463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-mvvcd" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.655470 4921 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-03T03:43:31.073192995Z","Handler":null,"Name":""} Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.659638 4921 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.659709 4921 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.722321 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.727876 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.824457 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.824975 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.825014 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.826354 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.827967 4921 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.828009 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.848566 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.864760 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.882989 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-mvvcd\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: W0103 03:43:31.922082 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0507148_3a4d_426d_83de_016ca35fa93f.slice/crio-b601c15c55551489b14e1147af0428325b2315632cb1033ebb2d62e0c69185ae WatchSource:0}: Error finding container b601c15c55551489b14e1147af0428325b2315632cb1033ebb2d62e0c69185ae: Status 404 returned error can't find the container with id b601c15c55551489b14e1147af0428325b2315632cb1033ebb2d62e0c69185ae Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.926572 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.926691 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.930684 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.930758 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.951825 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.987694 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:31 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:31 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:31 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:31 crc kubenswrapper[4921]: I0103 03:43:31.987780 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.113722 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.124059 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.132371 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.156875 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.163508 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.169609 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.183643 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.232123 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.232234 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.232439 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhfs5\" (UniqueName: \"kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.298490 4921 patch_prober.go:28] interesting pod/downloads-7954f5f757-gqd64 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.298521 4921 patch_prober.go:28] interesting pod/downloads-7954f5f757-gqd64 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.298557 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gqd64" podUID="9fa8138b-e464-4c9e-8db6-09e0f03e8507" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.298548 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-gqd64" podUID="9fa8138b-e464-4c9e-8db6-09e0f03e8507" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.335258 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.335380 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.335419 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhfs5\" (UniqueName: \"kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.336441 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.336671 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.364113 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhfs5\" (UniqueName: \"kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5\") pod \"redhat-operators-z668n\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.432900 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.432991 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.467592 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.495334 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.498222 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.498288 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.527228 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.556711 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.578113 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.579482 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.618338 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.641067 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e2a7a9f4707ff7074ce9dd7d75306c216ccb012a08f5086e6b863d2adb4524b1"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.643702 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.643816 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4sbq\" (UniqueName: \"kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.643852 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.699704 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" event={"ID":"8fa84952-f5e1-4a40-b263-31b8d6a8e361","Type":"ContainerStarted","Data":"a6d39a744024aba3fac0f050e4396056cc00902cf72972ebbbf75a2e75fdfead"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.721436 4921 generic.go:334] "Generic (PLEG): container finished" podID="4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" containerID="b2e39a43b31941ad6b4ac11ce08f580170f306366a0238d925cd36ec0a92f373" exitCode=0 Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.721510 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" event={"ID":"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c","Type":"ContainerDied","Data":"b2e39a43b31941ad6b4ac11ce08f580170f306366a0238d925cd36ec0a92f373"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.748676 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4sbq\" (UniqueName: \"kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.748730 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.748767 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.749764 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.750144 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.753445 4921 generic.go:334] "Generic (PLEG): container finished" podID="d0507148-3a4d-426d-83de-016ca35fa93f" containerID="6c43de89c426eaf3ac319e53a63df909ec451f14ed56736aaea17e3005e21cd0" exitCode=0 Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.753571 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerDied","Data":"6c43de89c426eaf3ac319e53a63df909ec451f14ed56736aaea17e3005e21cd0"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.753614 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerStarted","Data":"b601c15c55551489b14e1147af0428325b2315632cb1033ebb2d62e0c69185ae"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.790807 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4sbq\" (UniqueName: \"kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq\") pod \"redhat-operators-n8lrm\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.792642 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-k4ffk" podStartSLOduration=12.792558253 podStartE2EDuration="12.792558253s" podCreationTimestamp="2026-01-03 03:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:32.740508378 +0000 UTC m=+148.351935212" watchObservedRunningTime="2026-01-03 03:43:32.792558253 +0000 UTC m=+148.403985077" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.835930 4921 generic.go:334] "Generic (PLEG): container finished" podID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerID="f8c0afadc222a3b58cb5425ff75484f4b12618b6727f2d519356c173ccbd1d30" exitCode=0 Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.836045 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerDied","Data":"f8c0afadc222a3b58cb5425ff75484f4b12618b6727f2d519356c173ccbd1d30"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.836084 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerStarted","Data":"d110296cbd1a7409b1d96d72ed6446e215987f72708d7e920f1df145f4299dfe"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.843468 4921 generic.go:334] "Generic (PLEG): container finished" podID="63515f7b-2842-4bda-a846-6e12fb49aade" containerID="d483f8f3737904b6b789efec9b3704d919cabdc6b52c4cac6cb00f6746b637aa" exitCode=0 Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.844633 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerDied","Data":"d483f8f3737904b6b789efec9b3704d919cabdc6b52c4cac6cb00f6746b637aa"} Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.863537 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5vdkd" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.869588 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-znd66" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.914192 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.928764 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.929602 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.933794 4921 patch_prober.go:28] interesting pod/console-f9d7485db-2855s container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.933862 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2855s" podUID="2e3f92a2-ffe3-4ada-8f0f-f2854367b9e9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.955088 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:43:32 crc kubenswrapper[4921]: I0103 03:43:32.982300 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.017943 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:33 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:33 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:33 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.030017 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.188624 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:43:33 crc kubenswrapper[4921]: W0103 03:43:33.268519 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6184e585_30f0_42d3_996f_ff66dba239c5.slice/crio-45fa296cde242a933df2734848ef92f7de5250cc394eddbf2e0d7279429032ec WatchSource:0}: Error finding container 45fa296cde242a933df2734848ef92f7de5250cc394eddbf2e0d7279429032ec: Status 404 returned error can't find the container with id 45fa296cde242a933df2734848ef92f7de5250cc394eddbf2e0d7279429032ec Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.576819 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:43:33 crc kubenswrapper[4921]: W0103 03:43:33.592039 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fa04459_742b_4ab3_8910_f3db4335c7db.slice/crio-344e8036c9403ac60be24e4fab0e8f0d008ae988b6e7b35c4edf3acb035a8719 WatchSource:0}: Error finding container 344e8036c9403ac60be24e4fab0e8f0d008ae988b6e7b35c4edf3acb035a8719: Status 404 returned error can't find the container with id 344e8036c9403ac60be24e4fab0e8f0d008ae988b6e7b35c4edf3acb035a8719 Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.894552 4921 generic.go:334] "Generic (PLEG): container finished" podID="6184e585-30f0-42d3-996f-ff66dba239c5" containerID="6fcc382a23c32a92d08ae0b964139cd0cddfb9b28de2cf3287f7232dd631c3c3" exitCode=0 Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.895066 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerDied","Data":"6fcc382a23c32a92d08ae0b964139cd0cddfb9b28de2cf3287f7232dd631c3c3"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.895098 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerStarted","Data":"45fa296cde242a933df2734848ef92f7de5250cc394eddbf2e0d7279429032ec"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.898338 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e5d6400436901cf5e42996eef4eadcefe2e408ee80325ff35bf9d270a7328983"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.901720 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerStarted","Data":"344e8036c9403ac60be24e4fab0e8f0d008ae988b6e7b35c4edf3acb035a8719"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.907631 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" event={"ID":"c7d5610b-f8e0-4984-9907-ce33eb526161","Type":"ContainerStarted","Data":"9917ab2fceb8f7c406980ab44ac01e03f9068039069bcc4c71ab0e6ab6a99dfc"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.907678 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" event={"ID":"c7d5610b-f8e0-4984-9907-ce33eb526161","Type":"ContainerStarted","Data":"ea2800b0811a233029d5818631cee1b34a29f925ff5360aee493a3236a6150a1"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.907724 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.915830 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"874ccca74b2e41688c922a3e2aba87be677d85dcbc22aaa0b937b43b3e533538"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.915882 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f9221ce8d4e474c5a9a7bf021db1a032db457f0688a2e082e07475176a983a8a"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.916111 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.921437 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e8c2295a57eb1687ad94ade166bc186c9fa554875f53bac0f869bc9587c85697"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.921516 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"805f83cac6a65c8f9332c65c051861547e4057bebd8539b0044b4c76c92cdd49"} Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.986518 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:33 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:33 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:33 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:33 crc kubenswrapper[4921]: I0103 03:43:33.986580 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.069026 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" podStartSLOduration=130.068971585 podStartE2EDuration="2m10.068971585s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:43:34.04392318 +0000 UTC m=+149.655350004" watchObservedRunningTime="2026-01-03 03:43:34.068971585 +0000 UTC m=+149.680398409" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.483834 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.615920 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume\") pod \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.616054 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nfsp\" (UniqueName: \"kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp\") pod \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.616142 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume\") pod \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\" (UID: \"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c\") " Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.617840 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume" (OuterVolumeSpecName: "config-volume") pod "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" (UID: "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.628383 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp" (OuterVolumeSpecName: "kube-api-access-6nfsp") pod "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" (UID: "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c"). InnerVolumeSpecName "kube-api-access-6nfsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.630060 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" (UID: "4aad5cc0-feb3-45ed-b12d-a9cad9e3564c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.717955 4921 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-config-volume\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.718011 4921 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.718026 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nfsp\" (UniqueName: \"kubernetes.io/projected/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c-kube-api-access-6nfsp\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.930557 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" event={"ID":"4aad5cc0-feb3-45ed-b12d-a9cad9e3564c","Type":"ContainerDied","Data":"2b8ed6e0ac576916736107a464cfc0e665365cac1d4f8857774a1c313f4ffb49"} Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.930621 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b8ed6e0ac576916736107a464cfc0e665365cac1d4f8857774a1c313f4ffb49" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.930704 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87" Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.936657 4921 generic.go:334] "Generic (PLEG): container finished" podID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerID="ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75" exitCode=0 Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.936792 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerDied","Data":"ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75"} Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.986467 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:34 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:34 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:34 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:34 crc kubenswrapper[4921]: I0103 03:43:34.986539 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.261487 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 03 03:43:35 crc kubenswrapper[4921]: E0103 03:43:35.261884 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" containerName="collect-profiles" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.261904 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" containerName="collect-profiles" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.262056 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" containerName="collect-profiles" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.262611 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.265762 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.265945 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.270344 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.328449 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.328520 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.430715 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.430803 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.430956 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.453196 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.598904 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.990123 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:35 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:35 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:35 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:35 crc kubenswrapper[4921]: I0103 03:43:35.990851 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.032630 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.925410 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.926848 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.926953 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.930861 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.931117 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.976493 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.976546 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.986666 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a07a42-85fe-448d-babd-da838676515e","Type":"ContainerStarted","Data":"fde87e71a5c07e02b31e8b887ec477a9d54b4e3f77502105e9f34b70f824a33c"} Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.986729 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a07a42-85fe-448d-babd-da838676515e","Type":"ContainerStarted","Data":"19ed67375d6084bc57400de0f02c4c0fc4c668c0235cd6b0e30ec6e709fd121f"} Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.987185 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:36 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:36 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:36 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:36 crc kubenswrapper[4921]: I0103 03:43:36.987288 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.078029 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.078166 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.078354 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.160741 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.260288 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.673586 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.986905 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:37 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:37 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:37 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:37 crc kubenswrapper[4921]: I0103 03:43:37.987404 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:38 crc kubenswrapper[4921]: I0103 03:43:38.010436 4921 generic.go:334] "Generic (PLEG): container finished" podID="a2a07a42-85fe-448d-babd-da838676515e" containerID="fde87e71a5c07e02b31e8b887ec477a9d54b4e3f77502105e9f34b70f824a33c" exitCode=0 Jan 03 03:43:38 crc kubenswrapper[4921]: I0103 03:43:38.010574 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a07a42-85fe-448d-babd-da838676515e","Type":"ContainerDied","Data":"fde87e71a5c07e02b31e8b887ec477a9d54b4e3f77502105e9f34b70f824a33c"} Jan 03 03:43:38 crc kubenswrapper[4921]: I0103 03:43:38.038321 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"965aa66e-ae31-4953-96ff-2790c4f79a7c","Type":"ContainerStarted","Data":"ba2089fa0a2f262ffc3589be5e46cebbcc87bc21c9282806660d40078a6e3cda"} Jan 03 03:43:38 crc kubenswrapper[4921]: I0103 03:43:38.986721 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:38 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:38 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:38 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:38 crc kubenswrapper[4921]: I0103 03:43:38.986799 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.111999 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-kw6ql" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.546913 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.654061 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir\") pod \"a2a07a42-85fe-448d-babd-da838676515e\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.654196 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access\") pod \"a2a07a42-85fe-448d-babd-da838676515e\" (UID: \"a2a07a42-85fe-448d-babd-da838676515e\") " Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.654196 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a2a07a42-85fe-448d-babd-da838676515e" (UID: "a2a07a42-85fe-448d-babd-da838676515e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.654837 4921 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2a07a42-85fe-448d-babd-da838676515e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.663433 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a2a07a42-85fe-448d-babd-da838676515e" (UID: "a2a07a42-85fe-448d-babd-da838676515e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.756360 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2a07a42-85fe-448d-babd-da838676515e-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.990194 4921 patch_prober.go:28] interesting pod/router-default-5444994796-x8wkt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 03 03:43:39 crc kubenswrapper[4921]: [-]has-synced failed: reason withheld Jan 03 03:43:39 crc kubenswrapper[4921]: [+]process-running ok Jan 03 03:43:39 crc kubenswrapper[4921]: healthz check failed Jan 03 03:43:39 crc kubenswrapper[4921]: I0103 03:43:39.990295 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x8wkt" podUID="13f1ddbe-b27a-4db0-bd05-a37b298bdebd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.081337 4921 generic.go:334] "Generic (PLEG): container finished" podID="965aa66e-ae31-4953-96ff-2790c4f79a7c" containerID="52412663f097c69243555c10856cdf23a1a1cd66c19fd7f1bfc04ba9f2c7937d" exitCode=0 Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.081423 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"965aa66e-ae31-4953-96ff-2790c4f79a7c","Type":"ContainerDied","Data":"52412663f097c69243555c10856cdf23a1a1cd66c19fd7f1bfc04ba9f2c7937d"} Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.088841 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a2a07a42-85fe-448d-babd-da838676515e","Type":"ContainerDied","Data":"19ed67375d6084bc57400de0f02c4c0fc4c668c0235cd6b0e30ec6e709fd121f"} Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.088898 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19ed67375d6084bc57400de0f02c4c0fc4c668c0235cd6b0e30ec6e709fd121f" Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.088980 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.987094 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:40 crc kubenswrapper[4921]: I0103 03:43:40.992728 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-x8wkt" Jan 03 03:43:42 crc kubenswrapper[4921]: I0103 03:43:42.302437 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-gqd64" Jan 03 03:43:43 crc kubenswrapper[4921]: I0103 03:43:43.018903 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:43 crc kubenswrapper[4921]: I0103 03:43:43.023710 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2855s" Jan 03 03:43:46 crc kubenswrapper[4921]: I0103 03:43:46.429519 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:46 crc kubenswrapper[4921]: I0103 03:43:46.441516 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb5f46fd-64d0-4308-8fc6-5eff70d2521b-metrics-certs\") pod \"network-metrics-daemon-vsp88\" (UID: \"fb5f46fd-64d0-4308-8fc6-5eff70d2521b\") " pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:46 crc kubenswrapper[4921]: I0103 03:43:46.600331 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vsp88" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.072607 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.235891 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"965aa66e-ae31-4953-96ff-2790c4f79a7c","Type":"ContainerDied","Data":"ba2089fa0a2f262ffc3589be5e46cebbcc87bc21c9282806660d40078a6e3cda"} Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.235958 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba2089fa0a2f262ffc3589be5e46cebbcc87bc21c9282806660d40078a6e3cda" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.235970 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.265838 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access\") pod \"965aa66e-ae31-4953-96ff-2790c4f79a7c\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.266004 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir\") pod \"965aa66e-ae31-4953-96ff-2790c4f79a7c\" (UID: \"965aa66e-ae31-4953-96ff-2790c4f79a7c\") " Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.266418 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "965aa66e-ae31-4953-96ff-2790c4f79a7c" (UID: "965aa66e-ae31-4953-96ff-2790c4f79a7c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.303037 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "965aa66e-ae31-4953-96ff-2790c4f79a7c" (UID: "965aa66e-ae31-4953-96ff-2790c4f79a7c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.368529 4921 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/965aa66e-ae31-4953-96ff-2790c4f79a7c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:48 crc kubenswrapper[4921]: I0103 03:43:48.368589 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/965aa66e-ae31-4953-96ff-2790c4f79a7c-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:43:51 crc kubenswrapper[4921]: I0103 03:43:51.961262 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:44:00 crc kubenswrapper[4921]: E0103 03:44:00.509638 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 03 03:44:00 crc kubenswrapper[4921]: E0103 03:44:00.515798 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dgjzr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qh9w2_openshift-marketplace(d5545afd-0acb-4149-8e56-2804de448a28): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:00 crc kubenswrapper[4921]: E0103 03:44:00.519287 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qh9w2" podUID="d5545afd-0acb-4149-8e56-2804de448a28" Jan 03 03:44:01 crc kubenswrapper[4921]: I0103 03:44:01.175869 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:44:01 crc kubenswrapper[4921]: I0103 03:44:01.175953 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:44:03 crc kubenswrapper[4921]: I0103 03:44:03.633308 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-zzrtr" Jan 03 03:44:05 crc kubenswrapper[4921]: E0103 03:44:05.641847 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qh9w2" podUID="d5545afd-0acb-4149-8e56-2804de448a28" Jan 03 03:44:05 crc kubenswrapper[4921]: E0103 03:44:05.818835 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 03 03:44:05 crc kubenswrapper[4921]: E0103 03:44:05.819094 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s7hpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jj6h6_openshift-marketplace(6674474f-8b80-4e0a-a6a0-9b894d2b6add): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:05 crc kubenswrapper[4921]: E0103 03:44:05.820350 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jj6h6" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" Jan 03 03:44:06 crc kubenswrapper[4921]: E0103 03:44:06.294332 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 03 03:44:06 crc kubenswrapper[4921]: E0103 03:44:06.295387 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ltgqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5lkzl_openshift-marketplace(63515f7b-2842-4bda-a846-6e12fb49aade): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:06 crc kubenswrapper[4921]: E0103 03:44:06.296732 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5lkzl" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" Jan 03 03:44:10 crc kubenswrapper[4921]: E0103 03:44:10.004783 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jj6h6" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" Jan 03 03:44:10 crc kubenswrapper[4921]: E0103 03:44:10.006330 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5lkzl" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" Jan 03 03:44:10 crc kubenswrapper[4921]: E0103 03:44:10.088121 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 03 03:44:10 crc kubenswrapper[4921]: E0103 03:44:10.088835 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lhfs5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-z668n_openshift-marketplace(6184e585-30f0-42d3-996f-ff66dba239c5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:10 crc kubenswrapper[4921]: E0103 03:44:10.090027 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-z668n" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.456583 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-z668n" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.520478 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.520679 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kqs7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jwp9b_openshift-marketplace(d0507148-3a4d-426d-83de-016ca35fa93f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.521912 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jwp9b" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.550681 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.550921 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grqcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cpcqj_openshift-marketplace(a79c2918-942a-4a60-a300-e3981d43dd75): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.552412 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.552626 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v4sbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-n8lrm_openshift-marketplace(0fa04459-742b-4ab3-8910-f3db4335c7db): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.552732 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cpcqj" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.554745 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-n8lrm" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.564380 4921 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.564572 4921 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8rtrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-khcdk_openshift-marketplace(cb5cde82-e1a0-4d9f-800a-9aa70cc6e728): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 03 03:44:11 crc kubenswrapper[4921]: E0103 03:44:11.567306 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-khcdk" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" Jan 03 03:44:11 crc kubenswrapper[4921]: I0103 03:44:11.908620 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vsp88"] Jan 03 03:44:11 crc kubenswrapper[4921]: W0103 03:44:11.918891 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb5f46fd_64d0_4308_8fc6_5eff70d2521b.slice/crio-cdca13274473f89de68361d519f9a1d0f426b28e5a7bbb66053dfba82f6d45fc WatchSource:0}: Error finding container cdca13274473f89de68361d519f9a1d0f426b28e5a7bbb66053dfba82f6d45fc: Status 404 returned error can't find the container with id cdca13274473f89de68361d519f9a1d0f426b28e5a7bbb66053dfba82f6d45fc Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.129407 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.417571 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vsp88" event={"ID":"fb5f46fd-64d0-4308-8fc6-5eff70d2521b","Type":"ContainerStarted","Data":"cff4322686429b44fef0b58bba075cda9357a8394e5863426f77d2a553a497c4"} Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.420584 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vsp88" event={"ID":"fb5f46fd-64d0-4308-8fc6-5eff70d2521b","Type":"ContainerStarted","Data":"cdca13274473f89de68361d519f9a1d0f426b28e5a7bbb66053dfba82f6d45fc"} Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.420231 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-n8lrm" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.420316 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jwp9b" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.420491 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cpcqj" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.419504 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-khcdk" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.484337 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.484673 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="965aa66e-ae31-4953-96ff-2790c4f79a7c" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.484688 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="965aa66e-ae31-4953-96ff-2790c4f79a7c" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: E0103 03:44:12.484707 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2a07a42-85fe-448d-babd-da838676515e" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.484716 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2a07a42-85fe-448d-babd-da838676515e" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.484836 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="965aa66e-ae31-4953-96ff-2790c4f79a7c" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.484848 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2a07a42-85fe-448d-babd-da838676515e" containerName="pruner" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.485395 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.490495 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.492742 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.493025 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.593652 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.593750 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.695167 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.695909 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.696199 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.716254 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:12 crc kubenswrapper[4921]: I0103 03:44:12.811497 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:13 crc kubenswrapper[4921]: I0103 03:44:13.319397 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 03 03:44:13 crc kubenswrapper[4921]: W0103 03:44:13.332623 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod6b51b5f0_75f9_49f5_9534_adbb26842d79.slice/crio-4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4 WatchSource:0}: Error finding container 4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4: Status 404 returned error can't find the container with id 4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4 Jan 03 03:44:13 crc kubenswrapper[4921]: I0103 03:44:13.429375 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6b51b5f0-75f9-49f5-9534-adbb26842d79","Type":"ContainerStarted","Data":"4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4"} Jan 03 03:44:13 crc kubenswrapper[4921]: I0103 03:44:13.430811 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vsp88" event={"ID":"fb5f46fd-64d0-4308-8fc6-5eff70d2521b","Type":"ContainerStarted","Data":"a06e00f4b26e38018e31bf5ee01d86cb4c78f1cd5f6223ce63bca291d2c9d269"} Jan 03 03:44:13 crc kubenswrapper[4921]: I0103 03:44:13.451972 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vsp88" podStartSLOduration=169.451949127 podStartE2EDuration="2m49.451949127s" podCreationTimestamp="2026-01-03 03:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:44:13.447051421 +0000 UTC m=+189.058478245" watchObservedRunningTime="2026-01-03 03:44:13.451949127 +0000 UTC m=+189.063375951" Jan 03 03:44:14 crc kubenswrapper[4921]: I0103 03:44:14.440188 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6b51b5f0-75f9-49f5-9534-adbb26842d79","Type":"ContainerStarted","Data":"3a692b83eb44ddbd8ad2f2a13eb4cf49c10474759e5487bfe0e37a171ac02c90"} Jan 03 03:44:14 crc kubenswrapper[4921]: I0103 03:44:14.466968 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.466945312 podStartE2EDuration="2.466945312s" podCreationTimestamp="2026-01-03 03:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:44:14.460834943 +0000 UTC m=+190.072261767" watchObservedRunningTime="2026-01-03 03:44:14.466945312 +0000 UTC m=+190.078372136" Jan 03 03:44:15 crc kubenswrapper[4921]: I0103 03:44:15.449657 4921 generic.go:334] "Generic (PLEG): container finished" podID="6b51b5f0-75f9-49f5-9534-adbb26842d79" containerID="3a692b83eb44ddbd8ad2f2a13eb4cf49c10474759e5487bfe0e37a171ac02c90" exitCode=0 Jan 03 03:44:15 crc kubenswrapper[4921]: I0103 03:44:15.449890 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6b51b5f0-75f9-49f5-9534-adbb26842d79","Type":"ContainerDied","Data":"3a692b83eb44ddbd8ad2f2a13eb4cf49c10474759e5487bfe0e37a171ac02c90"} Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.749461 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.768352 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access\") pod \"6b51b5f0-75f9-49f5-9534-adbb26842d79\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.768402 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir\") pod \"6b51b5f0-75f9-49f5-9534-adbb26842d79\" (UID: \"6b51b5f0-75f9-49f5-9534-adbb26842d79\") " Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.768727 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6b51b5f0-75f9-49f5-9534-adbb26842d79" (UID: "6b51b5f0-75f9-49f5-9534-adbb26842d79"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.774880 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6b51b5f0-75f9-49f5-9534-adbb26842d79" (UID: "6b51b5f0-75f9-49f5-9534-adbb26842d79"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.869725 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6b51b5f0-75f9-49f5-9534-adbb26842d79-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:16 crc kubenswrapper[4921]: I0103 03:44:16.870226 4921 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6b51b5f0-75f9-49f5-9534-adbb26842d79-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:16 crc kubenswrapper[4921]: E0103 03:44:16.945332 4921 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-pod6b51b5f0_75f9_49f5_9534_adbb26842d79.slice\": RecentStats: unable to find data in memory cache]" Jan 03 03:44:17 crc kubenswrapper[4921]: I0103 03:44:17.469700 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"6b51b5f0-75f9-49f5-9534-adbb26842d79","Type":"ContainerDied","Data":"4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4"} Jan 03 03:44:17 crc kubenswrapper[4921]: I0103 03:44:17.469842 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e27664cf46cf9439f9ced732d38b5e09b6b0406befdf26bd4eaf3ac8e71c6f4" Jan 03 03:44:17 crc kubenswrapper[4921]: I0103 03:44:17.469842 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.678524 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 03 03:44:19 crc kubenswrapper[4921]: E0103 03:44:19.679295 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b51b5f0-75f9-49f5-9534-adbb26842d79" containerName="pruner" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.679313 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b51b5f0-75f9-49f5-9534-adbb26842d79" containerName="pruner" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.679469 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b51b5f0-75f9-49f5-9534-adbb26842d79" containerName="pruner" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.682334 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.686837 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.686936 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.687590 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.722113 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.722185 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.722207 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.823066 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.823126 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.823153 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.823188 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.823220 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:19 crc kubenswrapper[4921]: I0103 03:44:19.851200 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access\") pod \"installer-9-crc\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:20 crc kubenswrapper[4921]: I0103 03:44:20.042817 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:44:20 crc kubenswrapper[4921]: I0103 03:44:20.474797 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 03 03:44:20 crc kubenswrapper[4921]: W0103 03:44:20.484322 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9f8e382a_f2b1_4fa8_a741_a0dcc28f7445.slice/crio-b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e WatchSource:0}: Error finding container b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e: Status 404 returned error can't find the container with id b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e Jan 03 03:44:20 crc kubenswrapper[4921]: I0103 03:44:20.493249 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445","Type":"ContainerStarted","Data":"b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e"} Jan 03 03:44:21 crc kubenswrapper[4921]: I0103 03:44:21.502057 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445","Type":"ContainerStarted","Data":"c924de3e9e669d03b2363604989fceb78c2d4e4ba28381413adcfd1e2b866681"} Jan 03 03:44:21 crc kubenswrapper[4921]: I0103 03:44:21.551839 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.551801302 podStartE2EDuration="2.551801302s" podCreationTimestamp="2026-01-03 03:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:44:21.540813637 +0000 UTC m=+197.152240461" watchObservedRunningTime="2026-01-03 03:44:21.551801302 +0000 UTC m=+197.163228126" Jan 03 03:44:22 crc kubenswrapper[4921]: I0103 03:44:22.516788 4921 generic.go:334] "Generic (PLEG): container finished" podID="d5545afd-0acb-4149-8e56-2804de448a28" containerID="0181ebb9968161926091d16d487a9a85cba9819b4402adfe5c2dcc305a0f0d5b" exitCode=0 Jan 03 03:44:22 crc kubenswrapper[4921]: I0103 03:44:22.516868 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerDied","Data":"0181ebb9968161926091d16d487a9a85cba9819b4402adfe5c2dcc305a0f0d5b"} Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.524936 4921 generic.go:334] "Generic (PLEG): container finished" podID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerID="75c3d028166c131b673e9e268b1da5f87d168001a980eaf1e4be19a067c39d51" exitCode=0 Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.525042 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerDied","Data":"75c3d028166c131b673e9e268b1da5f87d168001a980eaf1e4be19a067c39d51"} Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.530607 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerStarted","Data":"9b06195fbbe2ff51f355597f1d1b8d19213b9a06f8c1fbaf4b7e0bda108e83a7"} Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.533921 4921 generic.go:334] "Generic (PLEG): container finished" podID="63515f7b-2842-4bda-a846-6e12fb49aade" containerID="cbd11d8e8881929c9fcc2bfa300f4903d0895c027f689921a8a66fc243d58c4c" exitCode=0 Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.534465 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerDied","Data":"cbd11d8e8881929c9fcc2bfa300f4903d0895c027f689921a8a66fc243d58c4c"} Jan 03 03:44:23 crc kubenswrapper[4921]: I0103 03:44:23.587632 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qh9w2" podStartSLOduration=4.081647498 podStartE2EDuration="55.587610085s" podCreationTimestamp="2026-01-03 03:43:28 +0000 UTC" firstStartedPulling="2026-01-03 03:43:31.592545041 +0000 UTC m=+147.203971865" lastFinishedPulling="2026-01-03 03:44:23.098507628 +0000 UTC m=+198.709934452" observedRunningTime="2026-01-03 03:44:23.583560612 +0000 UTC m=+199.194987446" watchObservedRunningTime="2026-01-03 03:44:23.587610085 +0000 UTC m=+199.199036909" Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.545796 4921 generic.go:334] "Generic (PLEG): container finished" podID="a79c2918-942a-4a60-a300-e3981d43dd75" containerID="b46abe1abc1975e41700481f715296771a1cb2729089f2d34781179674e03d82" exitCode=0 Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.545968 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerDied","Data":"b46abe1abc1975e41700481f715296771a1cb2729089f2d34781179674e03d82"} Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.549873 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerStarted","Data":"828eb13878805f6e8d371ef273a8950a0e7dce623d74d3801ce5528cc5574491"} Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.552101 4921 generic.go:334] "Generic (PLEG): container finished" podID="d0507148-3a4d-426d-83de-016ca35fa93f" containerID="0a4e7e0ee36b9a6c8c729f7cf147b196320407250abc0e63b307334cd6c93dc2" exitCode=0 Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.552138 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerDied","Data":"0a4e7e0ee36b9a6c8c729f7cf147b196320407250abc0e63b307334cd6c93dc2"} Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.554874 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerStarted","Data":"54fd1c950fcaf9118ff578fa028e2a25c2fb1c75b808544fd4a7eafb6ef8ca6a"} Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.586601 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5lkzl" podStartSLOduration=3.110735096 podStartE2EDuration="55.586580635s" podCreationTimestamp="2026-01-03 03:43:29 +0000 UTC" firstStartedPulling="2026-01-03 03:43:31.490784797 +0000 UTC m=+147.102211621" lastFinishedPulling="2026-01-03 03:44:23.966630336 +0000 UTC m=+199.578057160" observedRunningTime="2026-01-03 03:44:24.586190865 +0000 UTC m=+200.197617689" watchObservedRunningTime="2026-01-03 03:44:24.586580635 +0000 UTC m=+200.198007459" Jan 03 03:44:24 crc kubenswrapper[4921]: I0103 03:44:24.648343 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jj6h6" podStartSLOduration=4.243083179 podStartE2EDuration="56.648322499s" podCreationTimestamp="2026-01-03 03:43:28 +0000 UTC" firstStartedPulling="2026-01-03 03:43:31.53233814 +0000 UTC m=+147.143764964" lastFinishedPulling="2026-01-03 03:44:23.93757746 +0000 UTC m=+199.549004284" observedRunningTime="2026-01-03 03:44:24.646048306 +0000 UTC m=+200.257475150" watchObservedRunningTime="2026-01-03 03:44:24.648322499 +0000 UTC m=+200.259749323" Jan 03 03:44:25 crc kubenswrapper[4921]: I0103 03:44:25.572584 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerStarted","Data":"d229abaed56cbed3ed2aa2ae93931d5be2d2bab8e8f797c64d3aafef15456bf5"} Jan 03 03:44:25 crc kubenswrapper[4921]: I0103 03:44:25.574640 4921 generic.go:334] "Generic (PLEG): container finished" podID="6184e585-30f0-42d3-996f-ff66dba239c5" containerID="a1a7346f137c8744f39c26aa11957113707e346d1d5681f96c613bf2692ad8fa" exitCode=0 Jan 03 03:44:25 crc kubenswrapper[4921]: I0103 03:44:25.574697 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerDied","Data":"a1a7346f137c8744f39c26aa11957113707e346d1d5681f96c613bf2692ad8fa"} Jan 03 03:44:25 crc kubenswrapper[4921]: I0103 03:44:25.581331 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerStarted","Data":"6757a99ea41e31ddd4f936361572117d7be5881de75dffea33f5c04237632a79"} Jan 03 03:44:25 crc kubenswrapper[4921]: I0103 03:44:25.610389 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jwp9b" podStartSLOduration=2.154593473 podStartE2EDuration="54.610362085s" podCreationTimestamp="2026-01-03 03:43:31 +0000 UTC" firstStartedPulling="2026-01-03 03:43:32.806945023 +0000 UTC m=+148.418371847" lastFinishedPulling="2026-01-03 03:44:25.262713645 +0000 UTC m=+200.874140459" observedRunningTime="2026-01-03 03:44:25.608675888 +0000 UTC m=+201.220102712" watchObservedRunningTime="2026-01-03 03:44:25.610362085 +0000 UTC m=+201.221788909" Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.590194 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerStarted","Data":"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59"} Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.592957 4921 generic.go:334] "Generic (PLEG): container finished" podID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerID="3eabdcb5341d572a13d08f46410f97fd2a71fc6ea1da5153476a1d0e726f764f" exitCode=0 Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.593007 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerDied","Data":"3eabdcb5341d572a13d08f46410f97fd2a71fc6ea1da5153476a1d0e726f764f"} Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.597036 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerStarted","Data":"5d7c9ae78bdd293cdcc7cce232469d5a3a475c36d2d0c820ec432ef573c5154d"} Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.616748 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cpcqj" podStartSLOduration=3.950117047 podStartE2EDuration="57.616705861s" podCreationTimestamp="2026-01-03 03:43:29 +0000 UTC" firstStartedPulling="2026-01-03 03:43:31.490842708 +0000 UTC m=+147.102269532" lastFinishedPulling="2026-01-03 03:44:25.157431522 +0000 UTC m=+200.768858346" observedRunningTime="2026-01-03 03:44:25.660152777 +0000 UTC m=+201.271579591" watchObservedRunningTime="2026-01-03 03:44:26.616705861 +0000 UTC m=+202.228132685" Jan 03 03:44:26 crc kubenswrapper[4921]: I0103 03:44:26.650206 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z668n" podStartSLOduration=2.535988219 podStartE2EDuration="54.65017776s" podCreationTimestamp="2026-01-03 03:43:32 +0000 UTC" firstStartedPulling="2026-01-03 03:43:33.897022242 +0000 UTC m=+149.508449066" lastFinishedPulling="2026-01-03 03:44:26.011211783 +0000 UTC m=+201.622638607" observedRunningTime="2026-01-03 03:44:26.645190101 +0000 UTC m=+202.256616925" watchObservedRunningTime="2026-01-03 03:44:26.65017776 +0000 UTC m=+202.261604584" Jan 03 03:44:27 crc kubenswrapper[4921]: I0103 03:44:27.603732 4921 generic.go:334] "Generic (PLEG): container finished" podID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerID="2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59" exitCode=0 Jan 03 03:44:27 crc kubenswrapper[4921]: I0103 03:44:27.603769 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerDied","Data":"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59"} Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.198396 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.198455 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.358529 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.452508 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.452725 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.497485 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.659655 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.660098 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.676412 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.681975 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.732818 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.822676 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.822755 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:29 crc kubenswrapper[4921]: I0103 03:44:29.871197 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:30 crc kubenswrapper[4921]: I0103 03:44:30.631655 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerStarted","Data":"e0a67b499b6b7f6b5fa3fe44f7fcee9aad934fd1fabc5777d1212f39035a3ea0"} Jan 03 03:44:30 crc kubenswrapper[4921]: I0103 03:44:30.673234 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:30 crc kubenswrapper[4921]: I0103 03:44:30.677135 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.176389 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.176864 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.176930 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.177837 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.177974 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625" gracePeriod=600 Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.518031 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.534368 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.534457 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.584992 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:31 crc kubenswrapper[4921]: I0103 03:44:31.687156 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.497068 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.497164 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.561484 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.644999 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625" exitCode=0 Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.645159 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625"} Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.645984 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cpcqj" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="registry-server" containerID="cri-o://6757a99ea41e31ddd4f936361572117d7be5881de75dffea33f5c04237632a79" gracePeriod=2 Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.677304 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-khcdk" podStartSLOduration=8.190222538 podStartE2EDuration="1m2.677256497s" podCreationTimestamp="2026-01-03 03:43:30 +0000 UTC" firstStartedPulling="2026-01-03 03:43:32.837660445 +0000 UTC m=+148.449087269" lastFinishedPulling="2026-01-03 03:44:27.324694404 +0000 UTC m=+202.936121228" observedRunningTime="2026-01-03 03:44:32.676466695 +0000 UTC m=+208.287893519" watchObservedRunningTime="2026-01-03 03:44:32.677256497 +0000 UTC m=+208.288683331" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.696844 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.921661 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:44:32 crc kubenswrapper[4921]: I0103 03:44:32.922613 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5lkzl" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="registry-server" containerID="cri-o://54fd1c950fcaf9118ff578fa028e2a25c2fb1c75b808544fd4a7eafb6ef8ca6a" gracePeriod=2 Jan 03 03:44:33 crc kubenswrapper[4921]: I0103 03:44:33.918952 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:44:33 crc kubenswrapper[4921]: I0103 03:44:33.919294 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jwp9b" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="registry-server" containerID="cri-o://d229abaed56cbed3ed2aa2ae93931d5be2d2bab8e8f797c64d3aafef15456bf5" gracePeriod=2 Jan 03 03:44:34 crc kubenswrapper[4921]: I0103 03:44:34.664070 4921 generic.go:334] "Generic (PLEG): container finished" podID="63515f7b-2842-4bda-a846-6e12fb49aade" containerID="54fd1c950fcaf9118ff578fa028e2a25c2fb1c75b808544fd4a7eafb6ef8ca6a" exitCode=0 Jan 03 03:44:34 crc kubenswrapper[4921]: I0103 03:44:34.664164 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerDied","Data":"54fd1c950fcaf9118ff578fa028e2a25c2fb1c75b808544fd4a7eafb6ef8ca6a"} Jan 03 03:44:34 crc kubenswrapper[4921]: I0103 03:44:34.667208 4921 generic.go:334] "Generic (PLEG): container finished" podID="a79c2918-942a-4a60-a300-e3981d43dd75" containerID="6757a99ea41e31ddd4f936361572117d7be5881de75dffea33f5c04237632a79" exitCode=0 Jan 03 03:44:34 crc kubenswrapper[4921]: I0103 03:44:34.667248 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerDied","Data":"6757a99ea41e31ddd4f936361572117d7be5881de75dffea33f5c04237632a79"} Jan 03 03:44:35 crc kubenswrapper[4921]: I0103 03:44:35.680175 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329"} Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.049181 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.118710 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178046 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities\") pod \"63515f7b-2842-4bda-a846-6e12fb49aade\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178183 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content\") pod \"a79c2918-942a-4a60-a300-e3981d43dd75\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178242 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content\") pod \"63515f7b-2842-4bda-a846-6e12fb49aade\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178305 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltgqq\" (UniqueName: \"kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq\") pod \"63515f7b-2842-4bda-a846-6e12fb49aade\" (UID: \"63515f7b-2842-4bda-a846-6e12fb49aade\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178335 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities\") pod \"a79c2918-942a-4a60-a300-e3981d43dd75\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.178388 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grqcz\" (UniqueName: \"kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz\") pod \"a79c2918-942a-4a60-a300-e3981d43dd75\" (UID: \"a79c2918-942a-4a60-a300-e3981d43dd75\") " Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.179240 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities" (OuterVolumeSpecName: "utilities") pod "63515f7b-2842-4bda-a846-6e12fb49aade" (UID: "63515f7b-2842-4bda-a846-6e12fb49aade"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.179315 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities" (OuterVolumeSpecName: "utilities") pod "a79c2918-942a-4a60-a300-e3981d43dd75" (UID: "a79c2918-942a-4a60-a300-e3981d43dd75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.186176 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz" (OuterVolumeSpecName: "kube-api-access-grqcz") pod "a79c2918-942a-4a60-a300-e3981d43dd75" (UID: "a79c2918-942a-4a60-a300-e3981d43dd75"). InnerVolumeSpecName "kube-api-access-grqcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.197538 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq" (OuterVolumeSpecName: "kube-api-access-ltgqq") pod "63515f7b-2842-4bda-a846-6e12fb49aade" (UID: "63515f7b-2842-4bda-a846-6e12fb49aade"). InnerVolumeSpecName "kube-api-access-ltgqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.244885 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63515f7b-2842-4bda-a846-6e12fb49aade" (UID: "63515f7b-2842-4bda-a846-6e12fb49aade"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.280816 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.280898 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltgqq\" (UniqueName: \"kubernetes.io/projected/63515f7b-2842-4bda-a846-6e12fb49aade-kube-api-access-ltgqq\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.280918 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.280936 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grqcz\" (UniqueName: \"kubernetes.io/projected/a79c2918-942a-4a60-a300-e3981d43dd75-kube-api-access-grqcz\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.280954 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63515f7b-2842-4bda-a846-6e12fb49aade-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.705514 4921 generic.go:334] "Generic (PLEG): container finished" podID="d0507148-3a4d-426d-83de-016ca35fa93f" containerID="d229abaed56cbed3ed2aa2ae93931d5be2d2bab8e8f797c64d3aafef15456bf5" exitCode=0 Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.705618 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerDied","Data":"d229abaed56cbed3ed2aa2ae93931d5be2d2bab8e8f797c64d3aafef15456bf5"} Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.710187 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5lkzl" event={"ID":"63515f7b-2842-4bda-a846-6e12fb49aade","Type":"ContainerDied","Data":"ca2f733d699ff2d8e395df2fd037bb89a0a112d90799b9bdb349b59fe99e1f25"} Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.710210 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5lkzl" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.710250 4921 scope.go:117] "RemoveContainer" containerID="54fd1c950fcaf9118ff578fa028e2a25c2fb1c75b808544fd4a7eafb6ef8ca6a" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.713482 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cpcqj" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.713594 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cpcqj" event={"ID":"a79c2918-942a-4a60-a300-e3981d43dd75","Type":"ContainerDied","Data":"ce1a842c47a7c8e1e6ee9206c9362969a7cf4d03e515218ec114232ffa14b19a"} Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.752970 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a79c2918-942a-4a60-a300-e3981d43dd75" (UID: "a79c2918-942a-4a60-a300-e3981d43dd75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.755058 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.764487 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5lkzl"] Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.789422 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a79c2918-942a-4a60-a300-e3981d43dd75-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:36 crc kubenswrapper[4921]: I0103 03:44:36.891223 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" path="/var/lib/kubelet/pods/63515f7b-2842-4bda-a846-6e12fb49aade/volumes" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.037931 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.041098 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cpcqj"] Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.200865 4921 scope.go:117] "RemoveContainer" containerID="cbd11d8e8881929c9fcc2bfa300f4903d0895c027f689921a8a66fc243d58c4c" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.268325 4921 scope.go:117] "RemoveContainer" containerID="d483f8f3737904b6b789efec9b3704d919cabdc6b52c4cac6cb00f6746b637aa" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.296252 4921 scope.go:117] "RemoveContainer" containerID="6757a99ea41e31ddd4f936361572117d7be5881de75dffea33f5c04237632a79" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.326067 4921 scope.go:117] "RemoveContainer" containerID="b46abe1abc1975e41700481f715296771a1cb2729089f2d34781179674e03d82" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.367141 4921 scope.go:117] "RemoveContainer" containerID="72f9e0b7124e5eb3cb96a7a9713c397bb31d2df81e7e3a602ecf2e108ccb42bf" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.564247 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.700896 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kqs7\" (UniqueName: \"kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7\") pod \"d0507148-3a4d-426d-83de-016ca35fa93f\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.701005 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content\") pod \"d0507148-3a4d-426d-83de-016ca35fa93f\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.701146 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities\") pod \"d0507148-3a4d-426d-83de-016ca35fa93f\" (UID: \"d0507148-3a4d-426d-83de-016ca35fa93f\") " Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.702076 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities" (OuterVolumeSpecName: "utilities") pod "d0507148-3a4d-426d-83de-016ca35fa93f" (UID: "d0507148-3a4d-426d-83de-016ca35fa93f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.708150 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7" (OuterVolumeSpecName: "kube-api-access-5kqs7") pod "d0507148-3a4d-426d-83de-016ca35fa93f" (UID: "d0507148-3a4d-426d-83de-016ca35fa93f"). InnerVolumeSpecName "kube-api-access-5kqs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.723862 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0507148-3a4d-426d-83de-016ca35fa93f" (UID: "d0507148-3a4d-426d-83de-016ca35fa93f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.725744 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerStarted","Data":"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17"} Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.727911 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jwp9b" event={"ID":"d0507148-3a4d-426d-83de-016ca35fa93f","Type":"ContainerDied","Data":"b601c15c55551489b14e1147af0428325b2315632cb1033ebb2d62e0c69185ae"} Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.727944 4921 scope.go:117] "RemoveContainer" containerID="d229abaed56cbed3ed2aa2ae93931d5be2d2bab8e8f797c64d3aafef15456bf5" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.728053 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jwp9b" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.749193 4921 scope.go:117] "RemoveContainer" containerID="0a4e7e0ee36b9a6c8c729f7cf147b196320407250abc0e63b307334cd6c93dc2" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.751945 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n8lrm" podStartSLOduration=3.423550008 podStartE2EDuration="1m5.751921266s" podCreationTimestamp="2026-01-03 03:43:32 +0000 UTC" firstStartedPulling="2026-01-03 03:43:34.939982504 +0000 UTC m=+150.551409328" lastFinishedPulling="2026-01-03 03:44:37.268353762 +0000 UTC m=+212.879780586" observedRunningTime="2026-01-03 03:44:37.749669063 +0000 UTC m=+213.361095877" watchObservedRunningTime="2026-01-03 03:44:37.751921266 +0000 UTC m=+213.363348090" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.762875 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.765701 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jwp9b"] Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.776683 4921 scope.go:117] "RemoveContainer" containerID="6c43de89c426eaf3ac319e53a63df909ec451f14ed56736aaea17e3005e21cd0" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.802424 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.802457 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0507148-3a4d-426d-83de-016ca35fa93f-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:37 crc kubenswrapper[4921]: I0103 03:44:37.802469 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kqs7\" (UniqueName: \"kubernetes.io/projected/d0507148-3a4d-426d-83de-016ca35fa93f-kube-api-access-5kqs7\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:38 crc kubenswrapper[4921]: I0103 03:44:38.893489 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" path="/var/lib/kubelet/pods/a79c2918-942a-4a60-a300-e3981d43dd75/volumes" Jan 03 03:44:38 crc kubenswrapper[4921]: I0103 03:44:38.894740 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" path="/var/lib/kubelet/pods/d0507148-3a4d-426d-83de-016ca35fa93f/volumes" Jan 03 03:44:41 crc kubenswrapper[4921]: I0103 03:44:41.121566 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:44:41 crc kubenswrapper[4921]: I0103 03:44:41.121988 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:44:41 crc kubenswrapper[4921]: I0103 03:44:41.184392 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:44:41 crc kubenswrapper[4921]: I0103 03:44:41.814125 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:44:42 crc kubenswrapper[4921]: I0103 03:44:42.020950 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qr9zc"] Jan 03 03:44:42 crc kubenswrapper[4921]: I0103 03:44:42.956480 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:42 crc kubenswrapper[4921]: I0103 03:44:42.956987 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:44 crc kubenswrapper[4921]: I0103 03:44:44.007100 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n8lrm" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="registry-server" probeResult="failure" output=< Jan 03 03:44:44 crc kubenswrapper[4921]: timeout: failed to connect service ":50051" within 1s Jan 03 03:44:44 crc kubenswrapper[4921]: > Jan 03 03:44:52 crc kubenswrapper[4921]: I0103 03:44:52.998011 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:53 crc kubenswrapper[4921]: I0103 03:44:53.051507 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:53 crc kubenswrapper[4921]: I0103 03:44:53.921753 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:44:54 crc kubenswrapper[4921]: I0103 03:44:54.849584 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n8lrm" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="registry-server" containerID="cri-o://58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17" gracePeriod=2 Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.836561 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.860900 4921 generic.go:334] "Generic (PLEG): container finished" podID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerID="58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17" exitCode=0 Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.860984 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n8lrm" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.860990 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerDied","Data":"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17"} Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.861025 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n8lrm" event={"ID":"0fa04459-742b-4ab3-8910-f3db4335c7db","Type":"ContainerDied","Data":"344e8036c9403ac60be24e4fab0e8f0d008ae988b6e7b35c4edf3acb035a8719"} Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.861047 4921 scope.go:117] "RemoveContainer" containerID="58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.890398 4921 scope.go:117] "RemoveContainer" containerID="2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.890682 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities\") pod \"0fa04459-742b-4ab3-8910-f3db4335c7db\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.890819 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4sbq\" (UniqueName: \"kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq\") pod \"0fa04459-742b-4ab3-8910-f3db4335c7db\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.890880 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content\") pod \"0fa04459-742b-4ab3-8910-f3db4335c7db\" (UID: \"0fa04459-742b-4ab3-8910-f3db4335c7db\") " Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.892680 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities" (OuterVolumeSpecName: "utilities") pod "0fa04459-742b-4ab3-8910-f3db4335c7db" (UID: "0fa04459-742b-4ab3-8910-f3db4335c7db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.898687 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq" (OuterVolumeSpecName: "kube-api-access-v4sbq") pod "0fa04459-742b-4ab3-8910-f3db4335c7db" (UID: "0fa04459-742b-4ab3-8910-f3db4335c7db"). InnerVolumeSpecName "kube-api-access-v4sbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.910901 4921 scope.go:117] "RemoveContainer" containerID="ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.935903 4921 scope.go:117] "RemoveContainer" containerID="58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17" Jan 03 03:44:55 crc kubenswrapper[4921]: E0103 03:44:55.936516 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17\": container with ID starting with 58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17 not found: ID does not exist" containerID="58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.936578 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17"} err="failed to get container status \"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17\": rpc error: code = NotFound desc = could not find container \"58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17\": container with ID starting with 58e0ee9c6d1c0231b0846d1d5ccf68e5d5bdb967c0ac51894385b38d251f4a17 not found: ID does not exist" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.936620 4921 scope.go:117] "RemoveContainer" containerID="2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59" Jan 03 03:44:55 crc kubenswrapper[4921]: E0103 03:44:55.936981 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59\": container with ID starting with 2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59 not found: ID does not exist" containerID="2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.937027 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59"} err="failed to get container status \"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59\": rpc error: code = NotFound desc = could not find container \"2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59\": container with ID starting with 2c89256ea638198c9d707932a24d95953c1ccb530c51469185828dec51e0ba59 not found: ID does not exist" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.937057 4921 scope.go:117] "RemoveContainer" containerID="ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75" Jan 03 03:44:55 crc kubenswrapper[4921]: E0103 03:44:55.937387 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75\": container with ID starting with ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75 not found: ID does not exist" containerID="ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.937460 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75"} err="failed to get container status \"ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75\": rpc error: code = NotFound desc = could not find container \"ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75\": container with ID starting with ba5c137a38511abbd2118b4ea3f63bb6b9720f3ffa55b1a89569d66612360e75 not found: ID does not exist" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.993012 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4sbq\" (UniqueName: \"kubernetes.io/projected/0fa04459-742b-4ab3-8910-f3db4335c7db-kube-api-access-v4sbq\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:55 crc kubenswrapper[4921]: I0103 03:44:55.993074 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:56 crc kubenswrapper[4921]: I0103 03:44:56.017683 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fa04459-742b-4ab3-8910-f3db4335c7db" (UID: "0fa04459-742b-4ab3-8910-f3db4335c7db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:44:56 crc kubenswrapper[4921]: I0103 03:44:56.095300 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa04459-742b-4ab3-8910-f3db4335c7db-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:44:56 crc kubenswrapper[4921]: I0103 03:44:56.202039 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:44:56 crc kubenswrapper[4921]: I0103 03:44:56.205465 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n8lrm"] Jan 03 03:44:56 crc kubenswrapper[4921]: I0103 03:44:56.897863 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" path="/var/lib/kubelet/pods/0fa04459-742b-4ab3-8910-f3db4335c7db/volumes" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.019789 4921 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.020750 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.020770 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.020785 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.020975 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.020989 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021000 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021013 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021020 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021037 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021045 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021057 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021066 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021077 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021084 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021098 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021106 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021126 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021135 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021146 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021154 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021166 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021174 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="extract-content" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.021185 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021193 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="extract-utilities" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021345 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa04459-742b-4ab3-8910-f3db4335c7db" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021370 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="63515f7b-2842-4bda-a846-6e12fb49aade" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021389 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0507148-3a4d-426d-83de-016ca35fa93f" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021401 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79c2918-942a-4a60-a300-e3981d43dd75" containerName="registry-server" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.021903 4921 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022177 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022425 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30" gracePeriod=15 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022485 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5" gracePeriod=15 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022623 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954" gracePeriod=15 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022581 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02" gracePeriod=15 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.022638 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c" gracePeriod=15 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.024480 4921 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.024898 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.024945 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.024972 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.024990 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.025019 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025036 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.025055 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025074 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.025094 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025111 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.025135 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025152 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.025171 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025187 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025474 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025524 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025552 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025573 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025598 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.025621 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.043040 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.043135 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.045843 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.046117 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.048213 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.089347 4921 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149752 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149829 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149872 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149956 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149990 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150037 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150077 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.149961 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150417 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150500 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150253 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150190 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.150192 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.252684 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.252876 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.252970 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.253043 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.253126 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.253158 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.390757 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.429857 4921 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18871bc2252a5f4d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-03 03:44:59.427348301 +0000 UTC m=+235.038775165,LastTimestamp:2026-01-03 03:44:59.427348301 +0000 UTC m=+235.038775165,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.576510 4921 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.576615 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.899441 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"bf3d40cc8734e21d9b7a3fe5bb564a853d2c682ce29459d323c476058b75152f"} Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.899549 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"961910db98d4539211b293913968f2a15851b1b93cf015c4d501c4b263a564b1"} Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.901221 4921 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:44:59 crc kubenswrapper[4921]: E0103 03:44:59.901522 4921 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.902096 4921 generic.go:334] "Generic (PLEG): container finished" podID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" containerID="c924de3e9e669d03b2363604989fceb78c2d4e4ba28381413adcfd1e2b866681" exitCode=0 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.902208 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445","Type":"ContainerDied","Data":"c924de3e9e669d03b2363604989fceb78c2d4e4ba28381413adcfd1e2b866681"} Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.902918 4921 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.903388 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.906264 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.908692 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.910900 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c" exitCode=0 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.910961 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5" exitCode=0 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.910997 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02" exitCode=0 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.911021 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954" exitCode=2 Jan 03 03:44:59 crc kubenswrapper[4921]: I0103 03:44:59.911027 4921 scope.go:117] "RemoveContainer" containerID="2e81fc72749fa716b6876646c277665d5b1a0ffa45ee39a2d2e15f7a531c42cd" Jan 03 03:45:00 crc kubenswrapper[4921]: I0103 03:45:00.920423 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 03 03:45:00 crc kubenswrapper[4921]: E0103 03:45:00.958975 4921 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" volumeName="registry-storage" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.245839 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.247390 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.302236 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir\") pod \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.302383 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access\") pod \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.302437 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock\") pod \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\" (UID: \"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.302720 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock" (OuterVolumeSpecName: "var-lock") pod "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" (UID: "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.302764 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" (UID: "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.312333 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" (UID: "9f8e382a-f2b1-4fa8-a741-a0dcc28f7445"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.403963 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.403997 4921 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-var-lock\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.404009 4921 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9f8e382a-f2b1-4fa8-a741-a0dcc28f7445-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.408742 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.409766 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.410526 4921 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.411129 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.505740 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.505878 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.505975 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.506101 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.506508 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.506519 4921 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.506552 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.608329 4921 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.608393 4921 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.934866 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.934844 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9f8e382a-f2b1-4fa8-a741-a0dcc28f7445","Type":"ContainerDied","Data":"b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e"} Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.935076 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b29d19ed9ee6d4120e93490df0a3ec431d69a1ee70b61628bbf1c90d6a3e3b5e" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.940788 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.942009 4921 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30" exitCode=0 Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.942113 4921 scope.go:117] "RemoveContainer" containerID="9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.942161 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.966224 4921 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.967443 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.976726 4921 scope.go:117] "RemoveContainer" containerID="88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.976832 4921 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:01 crc kubenswrapper[4921]: I0103 03:45:01.977333 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.003733 4921 scope.go:117] "RemoveContainer" containerID="8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.024311 4921 scope.go:117] "RemoveContainer" containerID="a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.050612 4921 scope.go:117] "RemoveContainer" containerID="e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.071909 4921 scope.go:117] "RemoveContainer" containerID="74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.094299 4921 scope.go:117] "RemoveContainer" containerID="9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.095004 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\": container with ID starting with 9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c not found: ID does not exist" containerID="9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.095082 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c"} err="failed to get container status \"9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\": rpc error: code = NotFound desc = could not find container \"9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c\": container with ID starting with 9e548ba5ba86e3c162b430940d85e0ff7c09f2824e84555611f972085404a87c not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.095119 4921 scope.go:117] "RemoveContainer" containerID="88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.095743 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\": container with ID starting with 88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5 not found: ID does not exist" containerID="88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.095804 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5"} err="failed to get container status \"88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\": rpc error: code = NotFound desc = could not find container \"88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5\": container with ID starting with 88db5e23a91487f31aacdd54f7b017d79bd1b0b28c49660d092e68341755e0c5 not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.095877 4921 scope.go:117] "RemoveContainer" containerID="8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.096333 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\": container with ID starting with 8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02 not found: ID does not exist" containerID="8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.096361 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02"} err="failed to get container status \"8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\": rpc error: code = NotFound desc = could not find container \"8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02\": container with ID starting with 8c0b48494457110643e3e4acaf44915294fc9c2284053f96cb1d6bb5e3addc02 not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.096377 4921 scope.go:117] "RemoveContainer" containerID="a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.096664 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\": container with ID starting with a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954 not found: ID does not exist" containerID="a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.096687 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954"} err="failed to get container status \"a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\": rpc error: code = NotFound desc = could not find container \"a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954\": container with ID starting with a9f2b179a84d4d6d0e22f9dd79d27c5229f074d3270e5e8adc2bbcacf5e27954 not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.096700 4921 scope.go:117] "RemoveContainer" containerID="e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.096978 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\": container with ID starting with e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30 not found: ID does not exist" containerID="e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.097034 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30"} err="failed to get container status \"e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\": rpc error: code = NotFound desc = could not find container \"e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30\": container with ID starting with e6aa13c6b1c664db5a3f0f153b10aa45e06e8590bac1024a1a371e129b8faf30 not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.097072 4921 scope.go:117] "RemoveContainer" containerID="74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21" Jan 03 03:45:02 crc kubenswrapper[4921]: E0103 03:45:02.097440 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\": container with ID starting with 74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21 not found: ID does not exist" containerID="74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.097469 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21"} err="failed to get container status \"74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\": rpc error: code = NotFound desc = could not find container \"74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21\": container with ID starting with 74b24bfe14610043994bcd3964880858810cd5ff88da98ef6cc341fd1e640e21 not found: ID does not exist" Jan 03 03:45:02 crc kubenswrapper[4921]: I0103 03:45:02.893444 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 03 03:45:03 crc kubenswrapper[4921]: E0103 03:45:03.872155 4921 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.251:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18871bc2252a5f4d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-03 03:44:59.427348301 +0000 UTC m=+235.038775165,LastTimestamp:2026-01-03 03:44:59.427348301 +0000 UTC m=+235.038775165,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 03 03:45:04 crc kubenswrapper[4921]: I0103 03:45:04.887261 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:07 crc kubenswrapper[4921]: I0103 03:45:07.056417 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" containerName="oauth-openshift" containerID="cri-o://a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429" gracePeriod=15 Jan 03 03:45:07 crc kubenswrapper[4921]: E0103 03:45:07.608535 4921 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod935ca01a_42e9_4cfc_9084_c1ded2c63949.slice/crio-conmon-a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429.scope\": RecentStats: unable to find data in memory cache]" Jan 03 03:45:07 crc kubenswrapper[4921]: I0103 03:45:07.959464 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:45:07 crc kubenswrapper[4921]: I0103 03:45:07.960193 4921 status_manager.go:851] "Failed to get status for pod" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qr9zc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:07 crc kubenswrapper[4921]: I0103 03:45:07.960800 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006413 4921 generic.go:334] "Generic (PLEG): container finished" podID="935ca01a-42e9-4cfc-9084-c1ded2c63949" containerID="a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429" exitCode=0 Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006485 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" event={"ID":"935ca01a-42e9-4cfc-9084-c1ded2c63949","Type":"ContainerDied","Data":"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429"} Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006525 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" event={"ID":"935ca01a-42e9-4cfc-9084-c1ded2c63949","Type":"ContainerDied","Data":"66e60183ebc4668124806101a71bc5ab328391ef918e00b3f47aac277a88ad0c"} Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006544 4921 scope.go:117] "RemoveContainer" containerID="a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006577 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.004923 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006739 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006802 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006881 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr6k5\" (UniqueName: \"kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.006958 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007014 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007447 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007534 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007597 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007637 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007669 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007735 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007807 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007858 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007939 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error\") pod \"935ca01a-42e9-4cfc-9084-c1ded2c63949\" (UID: \"935ca01a-42e9-4cfc-9084-c1ded2c63949\") " Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.007956 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.008504 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.008715 4921 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.008746 4921 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/935ca01a-42e9-4cfc-9084-c1ded2c63949-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.009419 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.010124 4921 status_manager.go:851] "Failed to get status for pod" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qr9zc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.010538 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.010803 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.015574 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.015641 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5" (OuterVolumeSpecName: "kube-api-access-mr6k5") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "kube-api-access-mr6k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.017216 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.018506 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.019708 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.021711 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.022153 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.022392 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.023102 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "935ca01a-42e9-4cfc-9084-c1ded2c63949" (UID: "935ca01a-42e9-4cfc-9084-c1ded2c63949"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.101010 4921 scope.go:117] "RemoveContainer" containerID="a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.102187 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429\": container with ID starting with a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429 not found: ID does not exist" containerID="a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.102334 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429"} err="failed to get container status \"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429\": rpc error: code = NotFound desc = could not find container \"a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429\": container with ID starting with a0cd36924e7e1e6195c5297c9e34b9a45e74f0f5fece921b0a05962810cef429 not found: ID does not exist" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110063 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr6k5\" (UniqueName: \"kubernetes.io/projected/935ca01a-42e9-4cfc-9084-c1ded2c63949-kube-api-access-mr6k5\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110089 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110102 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110114 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110124 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110136 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110146 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110157 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110169 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110178 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110187 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.110197 4921 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/935ca01a-42e9-4cfc-9084-c1ded2c63949-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.293004 4921 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.293339 4921 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.295548 4921 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.296264 4921 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.296721 4921 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.296776 4921 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.297137 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="200ms" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.321430 4921 status_manager.go:851] "Failed to get status for pod" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qr9zc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: I0103 03:45:08.321817 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.498924 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="400ms" Jan 03 03:45:08 crc kubenswrapper[4921]: E0103 03:45:08.900394 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="800ms" Jan 03 03:45:09 crc kubenswrapper[4921]: E0103 03:45:09.701858 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="1.6s" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.883615 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.886561 4921 status_manager.go:851] "Failed to get status for pod" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qr9zc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.887130 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.908002 4921 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.908079 4921 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:10 crc kubenswrapper[4921]: E0103 03:45:10.908942 4921 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:10 crc kubenswrapper[4921]: I0103 03:45:10.910204 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:10 crc kubenswrapper[4921]: W0103 03:45:10.941834 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-1d2c5559513f9abbc738fa32f681c25cf7fdd9e5281dc94dffdd3ee1a6d58b59 WatchSource:0}: Error finding container 1d2c5559513f9abbc738fa32f681c25cf7fdd9e5281dc94dffdd3ee1a6d58b59: Status 404 returned error can't find the container with id 1d2c5559513f9abbc738fa32f681c25cf7fdd9e5281dc94dffdd3ee1a6d58b59 Jan 03 03:45:11 crc kubenswrapper[4921]: I0103 03:45:11.031037 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1d2c5559513f9abbc738fa32f681c25cf7fdd9e5281dc94dffdd3ee1a6d58b59"} Jan 03 03:45:11 crc kubenswrapper[4921]: E0103 03:45:11.302947 4921 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.251:6443: connect: connection refused" interval="3.2s" Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.041010 4921 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="f36ccc16f1c04549e1e022a1344f82600bf12defa6b5517600bb85fbb725a07c" exitCode=0 Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.041090 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"f36ccc16f1c04549e1e022a1344f82600bf12defa6b5517600bb85fbb725a07c"} Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.041397 4921 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.041431 4921 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.042127 4921 status_manager.go:851] "Failed to get status for pod" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" pod="openshift-authentication/oauth-openshift-558db77b4-qr9zc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qr9zc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:12 crc kubenswrapper[4921]: E0103 03:45:12.042191 4921 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:12 crc kubenswrapper[4921]: I0103 03:45:12.042973 4921 status_manager.go:851] "Failed to get status for pod" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.251:6443: connect: connection refused" Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.059806 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d11c2c5a20bae3306adcb959494289a25a25b39d6c73f78ca4bf01ac730bf07c"} Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.060319 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a2d9392452a9055c299e3e99747266a97ae7ee26cf73225ef44eed9f6a3c5bb2"} Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.060344 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f7c5d73c982ec8e4dd9bf7af14a3f51917fe1ad6912af2caae970dedb44613d4"} Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.064237 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.064297 4921 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4" exitCode=1 Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.064347 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4"} Jan 03 03:45:13 crc kubenswrapper[4921]: I0103 03:45:13.064951 4921 scope.go:117] "RemoveContainer" containerID="e81f36c3804455e80622ebdb700ec815bb5defcbc33701839e33fd152c41f2a4" Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.075850 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.076297 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"41a6598e312c3e12909f4fb14eaa05ede2743d7ccda3159581f493d60546aecd"} Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.080392 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9e0d2355de12d0f3bcd743ef15a4f8cd356087af0afbf77fe4b1857a90b57488"} Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.080447 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"99734a8e9ecd484d6ad8534387294585f43622aba7012a101e24333d7f9d20e3"} Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.080622 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.080792 4921 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:14 crc kubenswrapper[4921]: I0103 03:45:14.080827 4921 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:15 crc kubenswrapper[4921]: I0103 03:45:15.910618 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:15 crc kubenswrapper[4921]: I0103 03:45:15.910712 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:15 crc kubenswrapper[4921]: I0103 03:45:15.916744 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:19 crc kubenswrapper[4921]: I0103 03:45:19.098051 4921 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:19 crc kubenswrapper[4921]: I0103 03:45:19.174128 4921 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="67515e17-3132-4de2-b54d-b4f5455240c8" Jan 03 03:45:19 crc kubenswrapper[4921]: I0103 03:45:19.265603 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:45:20 crc kubenswrapper[4921]: I0103 03:45:20.122907 4921 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:20 crc kubenswrapper[4921]: I0103 03:45:20.122943 4921 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:20 crc kubenswrapper[4921]: I0103 03:45:20.127549 4921 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="67515e17-3132-4de2-b54d-b4f5455240c8" Jan 03 03:45:22 crc kubenswrapper[4921]: I0103 03:45:22.662718 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:45:22 crc kubenswrapper[4921]: I0103 03:45:22.669297 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:45:23 crc kubenswrapper[4921]: I0103 03:45:23.150254 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 03 03:45:29 crc kubenswrapper[4921]: I0103 03:45:29.675954 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.036672 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.212916 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.410416 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.471057 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.507138 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.724146 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.741211 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 03 03:45:30 crc kubenswrapper[4921]: I0103 03:45:30.961224 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.331763 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.426136 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.552516 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.567032 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.610864 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.654173 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.761807 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.883671 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 03 03:45:31 crc kubenswrapper[4921]: I0103 03:45:31.927748 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.312707 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.402423 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.466197 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.595134 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.708127 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.736237 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.766778 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 03 03:45:32 crc kubenswrapper[4921]: I0103 03:45:32.953932 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.222121 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.359480 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.616478 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.626583 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.661706 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.672869 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.718564 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.722244 4921 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 03 03:45:33 crc kubenswrapper[4921]: I0103 03:45:33.801300 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.056209 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.056616 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.086978 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.113075 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.131705 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.205794 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.207412 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.216773 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.275806 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.379886 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.596896 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.706067 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.858605 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.904125 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.915696 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.915758 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.934355 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 03 03:45:34 crc kubenswrapper[4921]: I0103 03:45:34.963653 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.020181 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.021211 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.023088 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.065118 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.074244 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.076492 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.124422 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.126223 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.170046 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.221190 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.238970 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.286305 4921 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.294374 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.313821 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.432419 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.531335 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.537429 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.605448 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.712195 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.718257 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.807153 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.811551 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.933184 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.978845 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 03 03:45:35 crc kubenswrapper[4921]: I0103 03:45:35.988306 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.127612 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.235984 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.334561 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.403520 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.500200 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.582085 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.610589 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.666671 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.700171 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.784256 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.807691 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.812000 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.857699 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.892408 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.923576 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 03 03:45:36 crc kubenswrapper[4921]: I0103 03:45:36.975206 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.007055 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.041891 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.115471 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.152875 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.180066 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.200924 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.250404 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.261370 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.280592 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.292112 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.307347 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.398051 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.418053 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.465437 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.483846 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.489924 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.498811 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.540518 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.581071 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.593286 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.698702 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.786443 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.840807 4921 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.846335 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.911706 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 03 03:45:37 crc kubenswrapper[4921]: I0103 03:45:37.967628 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.038099 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.073691 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.098293 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.139033 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.243330 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.339106 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.398525 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.451728 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.451808 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.540164 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.613196 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.628659 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.876235 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 03 03:45:38 crc kubenswrapper[4921]: I0103 03:45:38.931211 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.071009 4921 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.194859 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.234132 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.401308 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.401420 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.631879 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.762806 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.782911 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.934726 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.935865 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.937090 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 03 03:45:39 crc kubenswrapper[4921]: I0103 03:45:39.950515 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.138434 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.171441 4921 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.291828 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.323513 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.425305 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.439648 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.581699 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.653436 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.659619 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.688244 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.722150 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.817722 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.829037 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.833791 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.959155 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 03 03:45:40 crc kubenswrapper[4921]: I0103 03:45:40.961531 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.123613 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.155629 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.206876 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.241884 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.245090 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.284206 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.293749 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.366559 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.430418 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.468317 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.494454 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.543655 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.584520 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.623043 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.632371 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.662574 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.803125 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.847803 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.887385 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 03 03:45:41 crc kubenswrapper[4921]: I0103 03:45:41.890035 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.089214 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.167328 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.184359 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.301076 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.342850 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.411799 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.450182 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.458830 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.617317 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.639962 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.664421 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.715532 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.744373 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.753025 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.784872 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 03 03:45:42 crc kubenswrapper[4921]: I0103 03:45:42.916004 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.003991 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.258660 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.348460 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.405568 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.535183 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.560860 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.591977 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.746325 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 03 03:45:43 crc kubenswrapper[4921]: I0103 03:45:43.755965 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.013105 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.053705 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.120881 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.149432 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.150642 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.205589 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.223857 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.240016 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.478490 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.539569 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.631057 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.656608 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.742194 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.836507 4921 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.842322 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.855989 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-qr9zc"] Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856114 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p","openshift-authentication/oauth-openshift-ccc74cc7-7gjnc"] Jan 03 03:45:44 crc kubenswrapper[4921]: E0103 03:45:44.856545 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" containerName="installer" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856590 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" containerName="installer" Jan 03 03:45:44 crc kubenswrapper[4921]: E0103 03:45:44.856632 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" containerName="oauth-openshift" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856648 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" containerName="oauth-openshift" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856682 4921 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856720 4921 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f16567fe-14ec-443c-966a-78e2e77f48fe" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856847 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" containerName="oauth-openshift" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.856889 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f8e382a-f2b1-4fa8-a741-a0dcc28f7445" containerName="installer" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.858249 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.859136 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.864254 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.864370 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.864516 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.864671 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.865199 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.865836 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.866249 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.866668 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.867254 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.867617 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.867945 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.868447 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.868802 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.868814 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.868818 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871135 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871283 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871376 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnhlw\" (UniqueName: \"kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871552 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg485\" (UniqueName: \"kubernetes.io/projected/6172c0ac-9909-46c6-af98-a3bc1798977e-kube-api-access-gg485\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871600 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871625 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871672 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871694 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871714 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871772 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-dir\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871813 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871876 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871934 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871959 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.871992 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.872017 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.872041 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-policies\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.873835 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.874966 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.883424 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.902848 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.903631 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="935ca01a-42e9-4cfc-9084-c1ded2c63949" path="/var/lib/kubelet/pods/935ca01a-42e9-4cfc-9084-c1ded2c63949/volumes" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.911964 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.911911009 podStartE2EDuration="25.911911009s" podCreationTimestamp="2026-01-03 03:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:45:44.90435271 +0000 UTC m=+280.515779534" watchObservedRunningTime="2026-01-03 03:45:44.911911009 +0000 UTC m=+280.523337853" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.954343 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.974901 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.974955 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.974980 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975014 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-dir\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975034 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975071 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975090 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975107 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975135 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975152 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975170 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-policies\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975196 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975232 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975254 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnhlw\" (UniqueName: \"kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975317 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg485\" (UniqueName: \"kubernetes.io/projected/6172c0ac-9909-46c6-af98-a3bc1798977e-kube-api-access-gg485\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975339 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.975356 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.977078 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.977608 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-service-ca\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.995992 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-policies\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:44 crc kubenswrapper[4921]: I0103 03:45:44.999806 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.007911 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.010768 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.010864 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6172c0ac-9909-46c6-af98-a3bc1798977e-audit-dir\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.013216 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-login\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.015049 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-router-certs\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.015855 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.016524 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-template-error\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.016832 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.017904 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-session\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.018611 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.027161 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6172c0ac-9909-46c6-af98-a3bc1798977e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.027629 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.037942 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnhlw\" (UniqueName: \"kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw\") pod \"collect-profiles-29456865-b6v9p\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.050768 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg485\" (UniqueName: \"kubernetes.io/projected/6172c0ac-9909-46c6-af98-a3bc1798977e-kube-api-access-gg485\") pod \"oauth-openshift-ccc74cc7-7gjnc\" (UID: \"6172c0ac-9909-46c6-af98-a3bc1798977e\") " pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.184822 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.192676 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.249550 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.383434 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.473798 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.625861 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.627722 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.656786 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 03 03:45:45 crc kubenswrapper[4921]: I0103 03:45:45.702423 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 03 03:45:46 crc kubenswrapper[4921]: I0103 03:45:46.065153 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 03 03:45:46 crc kubenswrapper[4921]: I0103 03:45:46.435900 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 03 03:45:46 crc kubenswrapper[4921]: I0103 03:45:46.957440 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 03 03:45:47 crc kubenswrapper[4921]: I0103 03:45:47.003889 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 03 03:45:47 crc kubenswrapper[4921]: I0103 03:45:47.568491 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 03 03:45:47 crc kubenswrapper[4921]: I0103 03:45:47.695940 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 03 03:45:48 crc kubenswrapper[4921]: I0103 03:45:48.073737 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-ccc74cc7-7gjnc"] Jan 03 03:45:48 crc kubenswrapper[4921]: I0103 03:45:48.091810 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p"] Jan 03 03:45:48 crc kubenswrapper[4921]: I0103 03:45:48.294871 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p"] Jan 03 03:45:48 crc kubenswrapper[4921]: I0103 03:45:48.319224 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" event={"ID":"edcf3091-1e30-435e-8fe5-633cf912b51d","Type":"ContainerStarted","Data":"814e65ee4317199109bc8a50e6681999c9fdb9e10192330236f5443a138819ae"} Jan 03 03:45:48 crc kubenswrapper[4921]: I0103 03:45:48.414098 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-ccc74cc7-7gjnc"] Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.331009 4921 generic.go:334] "Generic (PLEG): container finished" podID="edcf3091-1e30-435e-8fe5-633cf912b51d" containerID="9b078d72a72f8b15bea1624bb6204a7990fc87102a3476a37b0e205259a26bc1" exitCode=0 Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.331111 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" event={"ID":"edcf3091-1e30-435e-8fe5-633cf912b51d","Type":"ContainerDied","Data":"9b078d72a72f8b15bea1624bb6204a7990fc87102a3476a37b0e205259a26bc1"} Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.334308 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" event={"ID":"6172c0ac-9909-46c6-af98-a3bc1798977e","Type":"ContainerStarted","Data":"0dc1cebae67c3564e18009bedcec487a944e098b7502adf90dc3a7d85aef2c0f"} Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.334598 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" event={"ID":"6172c0ac-9909-46c6-af98-a3bc1798977e","Type":"ContainerStarted","Data":"0dc243ff3d7a7bd1651fa195066b517bfdc7ad2a7894e87530efaaf9266a8181"} Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.334819 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.395093 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" podStartSLOduration=67.395057024 podStartE2EDuration="1m7.395057024s" podCreationTimestamp="2026-01-03 03:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:45:49.388245657 +0000 UTC m=+284.999672521" watchObservedRunningTime="2026-01-03 03:45:49.395057024 +0000 UTC m=+285.006483888" Jan 03 03:45:49 crc kubenswrapper[4921]: I0103 03:45:49.527294 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-ccc74cc7-7gjnc" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.691411 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.868256 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume\") pod \"edcf3091-1e30-435e-8fe5-633cf912b51d\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.868433 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnhlw\" (UniqueName: \"kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw\") pod \"edcf3091-1e30-435e-8fe5-633cf912b51d\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.868541 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume\") pod \"edcf3091-1e30-435e-8fe5-633cf912b51d\" (UID: \"edcf3091-1e30-435e-8fe5-633cf912b51d\") " Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.869073 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume" (OuterVolumeSpecName: "config-volume") pod "edcf3091-1e30-435e-8fe5-633cf912b51d" (UID: "edcf3091-1e30-435e-8fe5-633cf912b51d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.879569 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "edcf3091-1e30-435e-8fe5-633cf912b51d" (UID: "edcf3091-1e30-435e-8fe5-633cf912b51d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.881329 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw" (OuterVolumeSpecName: "kube-api-access-jnhlw") pod "edcf3091-1e30-435e-8fe5-633cf912b51d" (UID: "edcf3091-1e30-435e-8fe5-633cf912b51d"). InnerVolumeSpecName "kube-api-access-jnhlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.970314 4921 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/edcf3091-1e30-435e-8fe5-633cf912b51d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.970357 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnhlw\" (UniqueName: \"kubernetes.io/projected/edcf3091-1e30-435e-8fe5-633cf912b51d-kube-api-access-jnhlw\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:50 crc kubenswrapper[4921]: I0103 03:45:50.970369 4921 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/edcf3091-1e30-435e-8fe5-633cf912b51d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:51 crc kubenswrapper[4921]: I0103 03:45:51.352070 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" Jan 03 03:45:51 crc kubenswrapper[4921]: I0103 03:45:51.352055 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456865-b6v9p" event={"ID":"edcf3091-1e30-435e-8fe5-633cf912b51d","Type":"ContainerDied","Data":"814e65ee4317199109bc8a50e6681999c9fdb9e10192330236f5443a138819ae"} Jan 03 03:45:51 crc kubenswrapper[4921]: I0103 03:45:51.352169 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="814e65ee4317199109bc8a50e6681999c9fdb9e10192330236f5443a138819ae" Jan 03 03:45:53 crc kubenswrapper[4921]: I0103 03:45:53.118488 4921 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 03 03:45:53 crc kubenswrapper[4921]: I0103 03:45:53.118816 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://bf3d40cc8734e21d9b7a3fe5bb564a853d2c682ce29459d323c476058b75152f" gracePeriod=5 Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.405759 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.406527 4921 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="bf3d40cc8734e21d9b7a3fe5bb564a853d2c682ce29459d323c476058b75152f" exitCode=137 Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.742738 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.742871 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912604 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912722 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912750 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912810 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912859 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.912901 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.913359 4921 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.913400 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.913430 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.913490 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:58 crc kubenswrapper[4921]: I0103 03:45:58.926179 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.015010 4921 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.015227 4921 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.015238 4921 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.015246 4921 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.418477 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.419102 4921 scope.go:117] "RemoveContainer" containerID="bf3d40cc8734e21d9b7a3fe5bb564a853d2c682ce29459d323c476058b75152f" Jan 03 03:45:59 crc kubenswrapper[4921]: I0103 03:45:59.419180 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 03 03:46:00 crc kubenswrapper[4921]: I0103 03:46:00.890017 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 03 03:46:04 crc kubenswrapper[4921]: I0103 03:46:04.726318 4921 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.258001 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.259239 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" podUID="1d74e762-4edc-473e-b643-af99acf8f210" containerName="controller-manager" containerID="cri-o://6e4128e707bd486ad8e6b8d24afbe639e38f3b5bb4084f05a960af43925113c5" gracePeriod=30 Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.349403 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.349723 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" podUID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" containerName="route-controller-manager" containerID="cri-o://b0d67aa5d43d8f9f5d8a5d1c9c3a804873dc26d84cd55a2f8f0665e8e3d4f015" gracePeriod=30 Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.531356 4921 generic.go:334] "Generic (PLEG): container finished" podID="1d74e762-4edc-473e-b643-af99acf8f210" containerID="6e4128e707bd486ad8e6b8d24afbe639e38f3b5bb4084f05a960af43925113c5" exitCode=0 Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.531454 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" event={"ID":"1d74e762-4edc-473e-b643-af99acf8f210","Type":"ContainerDied","Data":"6e4128e707bd486ad8e6b8d24afbe639e38f3b5bb4084f05a960af43925113c5"} Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.541006 4921 generic.go:334] "Generic (PLEG): container finished" podID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" containerID="b0d67aa5d43d8f9f5d8a5d1c9c3a804873dc26d84cd55a2f8f0665e8e3d4f015" exitCode=0 Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.541059 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" event={"ID":"4c87f43b-b9e0-4c92-985e-3dbad79ff859","Type":"ContainerDied","Data":"b0d67aa5d43d8f9f5d8a5d1c9c3a804873dc26d84cd55a2f8f0665e8e3d4f015"} Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.837265 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.843538 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964311 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkgtv\" (UniqueName: \"kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv\") pod \"1d74e762-4edc-473e-b643-af99acf8f210\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964367 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g2zf\" (UniqueName: \"kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf\") pod \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964422 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert\") pod \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964464 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config\") pod \"1d74e762-4edc-473e-b643-af99acf8f210\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964521 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca\") pod \"1d74e762-4edc-473e-b643-af99acf8f210\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964554 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert\") pod \"1d74e762-4edc-473e-b643-af99acf8f210\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964575 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles\") pod \"1d74e762-4edc-473e-b643-af99acf8f210\" (UID: \"1d74e762-4edc-473e-b643-af99acf8f210\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964598 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca\") pod \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.964627 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config\") pod \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\" (UID: \"4c87f43b-b9e0-4c92-985e-3dbad79ff859\") " Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.965636 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca" (OuterVolumeSpecName: "client-ca") pod "1d74e762-4edc-473e-b643-af99acf8f210" (UID: "1d74e762-4edc-473e-b643-af99acf8f210"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.965684 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1d74e762-4edc-473e-b643-af99acf8f210" (UID: "1d74e762-4edc-473e-b643-af99acf8f210"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.965704 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config" (OuterVolumeSpecName: "config") pod "1d74e762-4edc-473e-b643-af99acf8f210" (UID: "1d74e762-4edc-473e-b643-af99acf8f210"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.966600 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca" (OuterVolumeSpecName: "client-ca") pod "4c87f43b-b9e0-4c92-985e-3dbad79ff859" (UID: "4c87f43b-b9e0-4c92-985e-3dbad79ff859"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.966815 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config" (OuterVolumeSpecName: "config") pod "4c87f43b-b9e0-4c92-985e-3dbad79ff859" (UID: "4c87f43b-b9e0-4c92-985e-3dbad79ff859"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.972185 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv" (OuterVolumeSpecName: "kube-api-access-tkgtv") pod "1d74e762-4edc-473e-b643-af99acf8f210" (UID: "1d74e762-4edc-473e-b643-af99acf8f210"). InnerVolumeSpecName "kube-api-access-tkgtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.972207 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1d74e762-4edc-473e-b643-af99acf8f210" (UID: "1d74e762-4edc-473e-b643-af99acf8f210"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.972262 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf" (OuterVolumeSpecName: "kube-api-access-8g2zf") pod "4c87f43b-b9e0-4c92-985e-3dbad79ff859" (UID: "4c87f43b-b9e0-4c92-985e-3dbad79ff859"). InnerVolumeSpecName "kube-api-access-8g2zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:46:14 crc kubenswrapper[4921]: I0103 03:46:14.975827 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4c87f43b-b9e0-4c92-985e-3dbad79ff859" (UID: "4c87f43b-b9e0-4c92-985e-3dbad79ff859"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.066327 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkgtv\" (UniqueName: \"kubernetes.io/projected/1d74e762-4edc-473e-b643-af99acf8f210-kube-api-access-tkgtv\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067574 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g2zf\" (UniqueName: \"kubernetes.io/projected/4c87f43b-b9e0-4c92-985e-3dbad79ff859-kube-api-access-8g2zf\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067631 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c87f43b-b9e0-4c92-985e-3dbad79ff859-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067658 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067679 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067715 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d74e762-4edc-473e-b643-af99acf8f210-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067732 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1d74e762-4edc-473e-b643-af99acf8f210-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067753 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.067774 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c87f43b-b9e0-4c92-985e-3dbad79ff859-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.553031 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.553019 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xvdtc" event={"ID":"1d74e762-4edc-473e-b643-af99acf8f210","Type":"ContainerDied","Data":"22b79e7c6e242bff4c37835d425cc6cc0c4e2f009c7a7c3009db1557ae867f79"} Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.553734 4921 scope.go:117] "RemoveContainer" containerID="6e4128e707bd486ad8e6b8d24afbe639e38f3b5bb4084f05a960af43925113c5" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.556156 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" event={"ID":"4c87f43b-b9e0-4c92-985e-3dbad79ff859","Type":"ContainerDied","Data":"b92e0bb01c864cde810ce32afea76fb7f5c1f3edd6e46d6ab4babb7b1efedb5e"} Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.556225 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.589220 4921 scope.go:117] "RemoveContainer" containerID="b0d67aa5d43d8f9f5d8a5d1c9c3a804873dc26d84cd55a2f8f0665e8e3d4f015" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.616129 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.624650 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cwfsp"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.628615 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.634260 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xvdtc"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.736969 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:46:15 crc kubenswrapper[4921]: E0103 03:46:15.737421 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" containerName="route-controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737448 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" containerName="route-controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: E0103 03:46:15.737470 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edcf3091-1e30-435e-8fe5-633cf912b51d" containerName="collect-profiles" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737478 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="edcf3091-1e30-435e-8fe5-633cf912b51d" containerName="collect-profiles" Jan 03 03:46:15 crc kubenswrapper[4921]: E0103 03:46:15.737494 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737502 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 03 03:46:15 crc kubenswrapper[4921]: E0103 03:46:15.737512 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d74e762-4edc-473e-b643-af99acf8f210" containerName="controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737520 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d74e762-4edc-473e-b643-af99acf8f210" containerName="controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737701 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737719 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d74e762-4edc-473e-b643-af99acf8f210" containerName="controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737727 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="edcf3091-1e30-435e-8fe5-633cf912b51d" containerName="collect-profiles" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.737737 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" containerName="route-controller-manager" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.738304 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.740926 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.741235 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.741400 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.742371 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.743497 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.743837 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.743961 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.745066 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.749159 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.755297 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.756082 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.756514 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.759403 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.759705 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.760451 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.760680 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.760922 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882483 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882541 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j528n\" (UniqueName: \"kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882576 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882605 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pb6z\" (UniqueName: \"kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882630 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882700 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882752 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882797 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.882842 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984259 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984337 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j528n\" (UniqueName: \"kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984416 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984440 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pb6z\" (UniqueName: \"kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984491 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984513 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984550 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984640 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.984659 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.987090 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.987799 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.988170 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.988208 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.988989 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:15 crc kubenswrapper[4921]: I0103 03:46:15.993309 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:15.999994 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.009774 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j528n\" (UniqueName: \"kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n\") pod \"controller-manager-748d8f84b-52g5r\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.010968 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pb6z\" (UniqueName: \"kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z\") pod \"route-controller-manager-569d9684fb-twghg\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.101038 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.108146 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.350029 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.390916 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:16 crc kubenswrapper[4921]: W0103 03:46:16.396765 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod424342c2_5252_4d6c_b5d2_a3c2c7e1292f.slice/crio-37ac97c21463af0a8e484346b92d704388118c1e043e2222fbc9278e7d6505cc WatchSource:0}: Error finding container 37ac97c21463af0a8e484346b92d704388118c1e043e2222fbc9278e7d6505cc: Status 404 returned error can't find the container with id 37ac97c21463af0a8e484346b92d704388118c1e043e2222fbc9278e7d6505cc Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.566227 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" event={"ID":"424342c2-5252-4d6c-b5d2-a3c2c7e1292f","Type":"ContainerStarted","Data":"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0"} Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.566350 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" event={"ID":"424342c2-5252-4d6c-b5d2-a3c2c7e1292f","Type":"ContainerStarted","Data":"37ac97c21463af0a8e484346b92d704388118c1e043e2222fbc9278e7d6505cc"} Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.566695 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.568115 4921 patch_prober.go:28] interesting pod/controller-manager-748d8f84b-52g5r container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.568167 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.582032 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" event={"ID":"a0813d48-b371-4503-baa6-25f5c21fe0ca","Type":"ContainerStarted","Data":"fe8d0db9bc48ca594c0337aeb2e97ea9b020087cc33eecc7edeb7ea4cdc45f50"} Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.582470 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" event={"ID":"a0813d48-b371-4503-baa6-25f5c21fe0ca","Type":"ContainerStarted","Data":"54bb86ad636a41c18b1058d13478160bf5619ef174a89b5a6e9cf5bb39fb5106"} Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.583784 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.585007 4921 patch_prober.go:28] interesting pod/route-controller-manager-569d9684fb-twghg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.585170 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.596076 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" podStartSLOduration=2.596047552 podStartE2EDuration="2.596047552s" podCreationTimestamp="2026-01-03 03:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:46:16.589044469 +0000 UTC m=+312.200471303" watchObservedRunningTime="2026-01-03 03:46:16.596047552 +0000 UTC m=+312.207474416" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.611108 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" podStartSLOduration=2.611087726 podStartE2EDuration="2.611087726s" podCreationTimestamp="2026-01-03 03:46:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:46:16.608953627 +0000 UTC m=+312.220380461" watchObservedRunningTime="2026-01-03 03:46:16.611087726 +0000 UTC m=+312.222514560" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.893004 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d74e762-4edc-473e-b643-af99acf8f210" path="/var/lib/kubelet/pods/1d74e762-4edc-473e-b643-af99acf8f210/volumes" Jan 03 03:46:16 crc kubenswrapper[4921]: I0103 03:46:16.894092 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c87f43b-b9e0-4c92-985e-3dbad79ff859" path="/var/lib/kubelet/pods/4c87f43b-b9e0-4c92-985e-3dbad79ff859/volumes" Jan 03 03:46:17 crc kubenswrapper[4921]: I0103 03:46:17.592464 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:17 crc kubenswrapper[4921]: I0103 03:46:17.596993 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.239817 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.240991 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerName="controller-manager" containerID="cri-o://ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0" gracePeriod=30 Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.702789 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.826240 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles\") pod \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.826410 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config\") pod \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.826468 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert\") pod \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.826511 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca\") pod \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.826557 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j528n\" (UniqueName: \"kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n\") pod \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\" (UID: \"424342c2-5252-4d6c-b5d2-a3c2c7e1292f\") " Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.827921 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "424342c2-5252-4d6c-b5d2-a3c2c7e1292f" (UID: "424342c2-5252-4d6c-b5d2-a3c2c7e1292f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.828493 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config" (OuterVolumeSpecName: "config") pod "424342c2-5252-4d6c-b5d2-a3c2c7e1292f" (UID: "424342c2-5252-4d6c-b5d2-a3c2c7e1292f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.828933 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca" (OuterVolumeSpecName: "client-ca") pod "424342c2-5252-4d6c-b5d2-a3c2c7e1292f" (UID: "424342c2-5252-4d6c-b5d2-a3c2c7e1292f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.835434 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n" (OuterVolumeSpecName: "kube-api-access-j528n") pod "424342c2-5252-4d6c-b5d2-a3c2c7e1292f" (UID: "424342c2-5252-4d6c-b5d2-a3c2c7e1292f"). InnerVolumeSpecName "kube-api-access-j528n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.837370 4921 generic.go:334] "Generic (PLEG): container finished" podID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerID="ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0" exitCode=0 Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.837441 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" event={"ID":"424342c2-5252-4d6c-b5d2-a3c2c7e1292f","Type":"ContainerDied","Data":"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0"} Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.837488 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" event={"ID":"424342c2-5252-4d6c-b5d2-a3c2c7e1292f","Type":"ContainerDied","Data":"37ac97c21463af0a8e484346b92d704388118c1e043e2222fbc9278e7d6505cc"} Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.837482 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-748d8f84b-52g5r" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.837515 4921 scope.go:117] "RemoveContainer" containerID="ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.838352 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "424342c2-5252-4d6c-b5d2-a3c2c7e1292f" (UID: "424342c2-5252-4d6c-b5d2-a3c2c7e1292f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.904937 4921 scope.go:117] "RemoveContainer" containerID="ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0" Jan 03 03:46:54 crc kubenswrapper[4921]: E0103 03:46:54.905588 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0\": container with ID starting with ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0 not found: ID does not exist" containerID="ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.905640 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0"} err="failed to get container status \"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0\": rpc error: code = NotFound desc = could not find container \"ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0\": container with ID starting with ea3981be2fb672296caaaa6641306e26574a9d42e2adc146b9ce52b502ade2b0 not found: ID does not exist" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.929121 4921 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.929157 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.929167 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.929176 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:54 crc kubenswrapper[4921]: I0103 03:46:54.929185 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j528n\" (UniqueName: \"kubernetes.io/projected/424342c2-5252-4d6c-b5d2-a3c2c7e1292f-kube-api-access-j528n\") on node \"crc\" DevicePath \"\"" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.163817 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.170987 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-748d8f84b-52g5r"] Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.760367 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r"] Jan 03 03:46:55 crc kubenswrapper[4921]: E0103 03:46:55.760753 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerName="controller-manager" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.760777 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerName="controller-manager" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.760994 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" containerName="controller-manager" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.761720 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.765685 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.766641 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.766702 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.766793 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.766955 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.769408 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.780157 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.783677 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r"] Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.942608 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-proxy-ca-bundles\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.942708 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-client-ca\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.942761 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-config\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.942798 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8njqr\" (UniqueName: \"kubernetes.io/projected/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-kube-api-access-8njqr\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:55 crc kubenswrapper[4921]: I0103 03:46:55.942926 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-serving-cert\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.044059 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-serving-cert\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.044432 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-proxy-ca-bundles\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.044522 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-client-ca\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.044617 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-config\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.044716 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8njqr\" (UniqueName: \"kubernetes.io/projected/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-kube-api-access-8njqr\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.045524 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-client-ca\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.045861 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-proxy-ca-bundles\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.046154 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-config\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.048425 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-serving-cert\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.061076 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8njqr\" (UniqueName: \"kubernetes.io/projected/2cf23b64-f2a2-48dc-a8c9-c591ea250f42-kube-api-access-8njqr\") pod \"controller-manager-6ffd7674cf-qtc6r\" (UID: \"2cf23b64-f2a2-48dc-a8c9-c591ea250f42\") " pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.092807 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.294329 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r"] Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.858977 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" event={"ID":"2cf23b64-f2a2-48dc-a8c9-c591ea250f42","Type":"ContainerStarted","Data":"5e5b1dc7f7d40f1e07bc17df7a192603ceee6a4150a2fc2ab06160f447c2debe"} Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.859403 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.859419 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" event={"ID":"2cf23b64-f2a2-48dc-a8c9-c591ea250f42","Type":"ContainerStarted","Data":"aef45d290c7e53046fd0647769637460ce14d6e26836ed0c61b81dcbc2be171d"} Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.866111 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.884410 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6ffd7674cf-qtc6r" podStartSLOduration=2.884389287 podStartE2EDuration="2.884389287s" podCreationTimestamp="2026-01-03 03:46:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:46:56.881003804 +0000 UTC m=+352.492430628" watchObservedRunningTime="2026-01-03 03:46:56.884389287 +0000 UTC m=+352.495816121" Jan 03 03:46:56 crc kubenswrapper[4921]: I0103 03:46:56.891861 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="424342c2-5252-4d6c-b5d2-a3c2c7e1292f" path="/var/lib/kubelet/pods/424342c2-5252-4d6c-b5d2-a3c2c7e1292f/volumes" Jan 03 03:47:00 crc kubenswrapper[4921]: I0103 03:47:00.896999 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pldd9"] Jan 03 03:47:00 crc kubenswrapper[4921]: I0103 03:47:00.898302 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:00 crc kubenswrapper[4921]: I0103 03:47:00.927156 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pldd9"] Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.009767 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-certificates\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.009969 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-tls\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010047 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a8bcb66-d11b-4163-8782-512d8825f58f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010175 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-trusted-ca\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010303 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a8bcb66-d11b-4163-8782-512d8825f58f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010394 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-bound-sa-token\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010490 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.010574 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwd5f\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-kube-api-access-vwd5f\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.066125 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111356 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-tls\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111408 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a8bcb66-d11b-4163-8782-512d8825f58f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111431 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-trusted-ca\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111475 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a8bcb66-d11b-4163-8782-512d8825f58f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111519 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-bound-sa-token\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111541 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwd5f\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-kube-api-access-vwd5f\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.111573 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-certificates\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.112417 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6a8bcb66-d11b-4163-8782-512d8825f58f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.112779 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-certificates\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.113516 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a8bcb66-d11b-4163-8782-512d8825f58f-trusted-ca\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.118141 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6a8bcb66-d11b-4163-8782-512d8825f58f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.118726 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-registry-tls\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.130961 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwd5f\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-kube-api-access-vwd5f\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.140623 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a8bcb66-d11b-4163-8782-512d8825f58f-bound-sa-token\") pod \"image-registry-66df7c8f76-pldd9\" (UID: \"6a8bcb66-d11b-4163-8782-512d8825f58f\") " pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.176195 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.176258 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.217216 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.726125 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-pldd9"] Jan 03 03:47:01 crc kubenswrapper[4921]: I0103 03:47:01.895230 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" event={"ID":"6a8bcb66-d11b-4163-8782-512d8825f58f","Type":"ContainerStarted","Data":"dddeabf412a3df823dcd8cb9f5377112fe7f01df4128553fb966f94a2b63fe2f"} Jan 03 03:47:02 crc kubenswrapper[4921]: I0103 03:47:02.903585 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" event={"ID":"6a8bcb66-d11b-4163-8782-512d8825f58f","Type":"ContainerStarted","Data":"606391892c73f0c104e859739fd1c25ca72a9d0e3a47dc60bdeb3984b53a39a3"} Jan 03 03:47:02 crc kubenswrapper[4921]: I0103 03:47:02.904558 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:02 crc kubenswrapper[4921]: I0103 03:47:02.947215 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" podStartSLOduration=2.947185307 podStartE2EDuration="2.947185307s" podCreationTimestamp="2026-01-03 03:47:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:47:02.941525052 +0000 UTC m=+358.552951916" watchObservedRunningTime="2026-01-03 03:47:02.947185307 +0000 UTC m=+358.558612171" Jan 03 03:47:14 crc kubenswrapper[4921]: I0103 03:47:14.221830 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:47:14 crc kubenswrapper[4921]: I0103 03:47:14.222571 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerName="route-controller-manager" containerID="cri-o://fe8d0db9bc48ca594c0337aeb2e97ea9b020087cc33eecc7edeb7ea4cdc45f50" gracePeriod=30 Jan 03 03:47:14 crc kubenswrapper[4921]: I0103 03:47:14.987087 4921 generic.go:334] "Generic (PLEG): container finished" podID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerID="fe8d0db9bc48ca594c0337aeb2e97ea9b020087cc33eecc7edeb7ea4cdc45f50" exitCode=0 Jan 03 03:47:14 crc kubenswrapper[4921]: I0103 03:47:14.987206 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" event={"ID":"a0813d48-b371-4503-baa6-25f5c21fe0ca","Type":"ContainerDied","Data":"fe8d0db9bc48ca594c0337aeb2e97ea9b020087cc33eecc7edeb7ea4cdc45f50"} Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.241150 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.269992 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9"] Jan 03 03:47:15 crc kubenswrapper[4921]: E0103 03:47:15.270479 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerName="route-controller-manager" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.270493 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerName="route-controller-manager" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.270589 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" containerName="route-controller-manager" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.270986 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.282125 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9"] Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.363803 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pb6z\" (UniqueName: \"kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z\") pod \"a0813d48-b371-4503-baa6-25f5c21fe0ca\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.363996 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca\") pod \"a0813d48-b371-4503-baa6-25f5c21fe0ca\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.364040 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config\") pod \"a0813d48-b371-4503-baa6-25f5c21fe0ca\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.364076 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert\") pod \"a0813d48-b371-4503-baa6-25f5c21fe0ca\" (UID: \"a0813d48-b371-4503-baa6-25f5c21fe0ca\") " Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.364845 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca" (OuterVolumeSpecName: "client-ca") pod "a0813d48-b371-4503-baa6-25f5c21fe0ca" (UID: "a0813d48-b371-4503-baa6-25f5c21fe0ca"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.365003 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config" (OuterVolumeSpecName: "config") pod "a0813d48-b371-4503-baa6-25f5c21fe0ca" (UID: "a0813d48-b371-4503-baa6-25f5c21fe0ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.371080 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z" (OuterVolumeSpecName: "kube-api-access-6pb6z") pod "a0813d48-b371-4503-baa6-25f5c21fe0ca" (UID: "a0813d48-b371-4503-baa6-25f5c21fe0ca"). InnerVolumeSpecName "kube-api-access-6pb6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.371773 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a0813d48-b371-4503-baa6-25f5c21fe0ca" (UID: "a0813d48-b371-4503-baa6-25f5c21fe0ca"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.465818 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jj8p\" (UniqueName: \"kubernetes.io/projected/787be437-4269-4f4f-95c7-f4499f4a2e5a-kube-api-access-6jj8p\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.465950 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-config\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466019 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-client-ca\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466063 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/787be437-4269-4f4f-95c7-f4499f4a2e5a-serving-cert\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466173 4921 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-client-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466193 4921 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0813d48-b371-4503-baa6-25f5c21fe0ca-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466209 4921 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0813d48-b371-4503-baa6-25f5c21fe0ca-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.466225 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pb6z\" (UniqueName: \"kubernetes.io/projected/a0813d48-b371-4503-baa6-25f5c21fe0ca-kube-api-access-6pb6z\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.567331 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-config\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.567396 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-client-ca\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.567423 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/787be437-4269-4f4f-95c7-f4499f4a2e5a-serving-cert\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.567457 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jj8p\" (UniqueName: \"kubernetes.io/projected/787be437-4269-4f4f-95c7-f4499f4a2e5a-kube-api-access-6jj8p\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.568931 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-config\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.569476 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/787be437-4269-4f4f-95c7-f4499f4a2e5a-client-ca\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.574160 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/787be437-4269-4f4f-95c7-f4499f4a2e5a-serving-cert\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.600079 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jj8p\" (UniqueName: \"kubernetes.io/projected/787be437-4269-4f4f-95c7-f4499f4a2e5a-kube-api-access-6jj8p\") pod \"route-controller-manager-5d8858b84f-lp7p9\" (UID: \"787be437-4269-4f4f-95c7-f4499f4a2e5a\") " pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.898740 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.997573 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" event={"ID":"a0813d48-b371-4503-baa6-25f5c21fe0ca","Type":"ContainerDied","Data":"54bb86ad636a41c18b1058d13478160bf5619ef174a89b5a6e9cf5bb39fb5106"} Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.998031 4921 scope.go:117] "RemoveContainer" containerID="fe8d0db9bc48ca594c0337aeb2e97ea9b020087cc33eecc7edeb7ea4cdc45f50" Jan 03 03:47:15 crc kubenswrapper[4921]: I0103 03:47:15.997709 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg" Jan 03 03:47:16 crc kubenswrapper[4921]: I0103 03:47:16.054451 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:47:16 crc kubenswrapper[4921]: I0103 03:47:16.063972 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-569d9684fb-twghg"] Jan 03 03:47:16 crc kubenswrapper[4921]: W0103 03:47:16.413418 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod787be437_4269_4f4f_95c7_f4499f4a2e5a.slice/crio-c5b8f1a62d1388273f61c50ad0ed091f670b485652e7339f740713460040bb7f WatchSource:0}: Error finding container c5b8f1a62d1388273f61c50ad0ed091f670b485652e7339f740713460040bb7f: Status 404 returned error can't find the container with id c5b8f1a62d1388273f61c50ad0ed091f670b485652e7339f740713460040bb7f Jan 03 03:47:16 crc kubenswrapper[4921]: I0103 03:47:16.415982 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9"] Jan 03 03:47:16 crc kubenswrapper[4921]: I0103 03:47:16.893416 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0813d48-b371-4503-baa6-25f5c21fe0ca" path="/var/lib/kubelet/pods/a0813d48-b371-4503-baa6-25f5c21fe0ca/volumes" Jan 03 03:47:17 crc kubenswrapper[4921]: I0103 03:47:17.004966 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" event={"ID":"787be437-4269-4f4f-95c7-f4499f4a2e5a","Type":"ContainerStarted","Data":"c20ab5f3634345ca95ba53e2eaf0b55dc37776253e7b7ad5c53508417011f0c1"} Jan 03 03:47:17 crc kubenswrapper[4921]: I0103 03:47:17.005018 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" event={"ID":"787be437-4269-4f4f-95c7-f4499f4a2e5a","Type":"ContainerStarted","Data":"c5b8f1a62d1388273f61c50ad0ed091f670b485652e7339f740713460040bb7f"} Jan 03 03:47:17 crc kubenswrapper[4921]: I0103 03:47:17.005169 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:17 crc kubenswrapper[4921]: I0103 03:47:17.028138 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" podStartSLOduration=3.028120664 podStartE2EDuration="3.028120664s" podCreationTimestamp="2026-01-03 03:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:47:17.026305313 +0000 UTC m=+372.637732137" watchObservedRunningTime="2026-01-03 03:47:17.028120664 +0000 UTC m=+372.639547488" Jan 03 03:47:17 crc kubenswrapper[4921]: I0103 03:47:17.185128 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5d8858b84f-lp7p9" Jan 03 03:47:21 crc kubenswrapper[4921]: I0103 03:47:21.222924 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-pldd9" Jan 03 03:47:21 crc kubenswrapper[4921]: I0103 03:47:21.286652 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.594105 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.595075 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jj6h6" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="registry-server" containerID="cri-o://828eb13878805f6e8d371ef273a8950a0e7dce623d74d3801ce5528cc5574491" gracePeriod=30 Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.616416 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.616728 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qh9w2" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="registry-server" containerID="cri-o://9b06195fbbe2ff51f355597f1d1b8d19213b9a06f8c1fbaf4b7e0bda108e83a7" gracePeriod=30 Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.622706 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.623149 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerName="marketplace-operator" containerID="cri-o://e4b3743478d8831b4d8672f7e93d858e303d848e0f894b87dc8f48e5cfc04bf8" gracePeriod=30 Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.634126 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.634438 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-khcdk" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="registry-server" containerID="cri-o://e0a67b499b6b7f6b5fa3fe44f7fcee9aad934fd1fabc5777d1212f39035a3ea0" gracePeriod=30 Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.648559 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8s84l"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.649554 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.654368 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.654665 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z668n" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="registry-server" containerID="cri-o://5d7c9ae78bdd293cdcc7cce232469d5a3a475c36d2d0c820ec432ef573c5154d" gracePeriod=30 Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.666633 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8s84l"] Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.820162 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.820238 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.820262 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcknh\" (UniqueName: \"kubernetes.io/projected/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-kube-api-access-jcknh\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.921141 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.921228 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.921251 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcknh\" (UniqueName: \"kubernetes.io/projected/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-kube-api-access-jcknh\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.924362 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.930252 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:24 crc kubenswrapper[4921]: I0103 03:47:24.939955 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcknh\" (UniqueName: \"kubernetes.io/projected/549a7ea1-10f1-4e90-a2e3-9ef968ce88da-kube-api-access-jcknh\") pod \"marketplace-operator-79b997595-8s84l\" (UID: \"549a7ea1-10f1-4e90-a2e3-9ef968ce88da\") " pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.062490 4921 generic.go:334] "Generic (PLEG): container finished" podID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerID="e0a67b499b6b7f6b5fa3fe44f7fcee9aad934fd1fabc5777d1212f39035a3ea0" exitCode=0 Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.062781 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerDied","Data":"e0a67b499b6b7f6b5fa3fe44f7fcee9aad934fd1fabc5777d1212f39035a3ea0"} Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.065066 4921 generic.go:334] "Generic (PLEG): container finished" podID="6184e585-30f0-42d3-996f-ff66dba239c5" containerID="5d7c9ae78bdd293cdcc7cce232469d5a3a475c36d2d0c820ec432ef573c5154d" exitCode=0 Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.065107 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerDied","Data":"5d7c9ae78bdd293cdcc7cce232469d5a3a475c36d2d0c820ec432ef573c5154d"} Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.066883 4921 generic.go:334] "Generic (PLEG): container finished" podID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerID="e4b3743478d8831b4d8672f7e93d858e303d848e0f894b87dc8f48e5cfc04bf8" exitCode=0 Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.066924 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" event={"ID":"8c06a2f0-3608-4b31-b52b-db33781f3dfa","Type":"ContainerDied","Data":"e4b3743478d8831b4d8672f7e93d858e303d848e0f894b87dc8f48e5cfc04bf8"} Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.070922 4921 generic.go:334] "Generic (PLEG): container finished" podID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerID="828eb13878805f6e8d371ef273a8950a0e7dce623d74d3801ce5528cc5574491" exitCode=0 Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.070993 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerDied","Data":"828eb13878805f6e8d371ef273a8950a0e7dce623d74d3801ce5528cc5574491"} Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.073554 4921 generic.go:334] "Generic (PLEG): container finished" podID="d5545afd-0acb-4149-8e56-2804de448a28" containerID="9b06195fbbe2ff51f355597f1d1b8d19213b9a06f8c1fbaf4b7e0bda108e83a7" exitCode=0 Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.073584 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerDied","Data":"9b06195fbbe2ff51f355597f1d1b8d19213b9a06f8c1fbaf4b7e0bda108e83a7"} Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.107113 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.197674 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.332237 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.335848 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content\") pod \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.336200 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities\") pod \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.336473 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7hpf\" (UniqueName: \"kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf\") pod \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\" (UID: \"6674474f-8b80-4e0a-a6a0-9b894d2b6add\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.339052 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities" (OuterVolumeSpecName: "utilities") pod "6674474f-8b80-4e0a-a6a0-9b894d2b6add" (UID: "6674474f-8b80-4e0a-a6a0-9b894d2b6add"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.342753 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.343717 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf" (OuterVolumeSpecName: "kube-api-access-s7hpf") pod "6674474f-8b80-4e0a-a6a0-9b894d2b6add" (UID: "6674474f-8b80-4e0a-a6a0-9b894d2b6add"). InnerVolumeSpecName "kube-api-access-s7hpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.391210 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.419197 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6674474f-8b80-4e0a-a6a0-9b894d2b6add" (UID: "6674474f-8b80-4e0a-a6a0-9b894d2b6add"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437599 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgjzr\" (UniqueName: \"kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr\") pod \"d5545afd-0acb-4149-8e56-2804de448a28\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437648 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics\") pod \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437678 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content\") pod \"d5545afd-0acb-4149-8e56-2804de448a28\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437705 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lvsd\" (UniqueName: \"kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd\") pod \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437741 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities\") pod \"d5545afd-0acb-4149-8e56-2804de448a28\" (UID: \"d5545afd-0acb-4149-8e56-2804de448a28\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.437816 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca\") pod \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\" (UID: \"8c06a2f0-3608-4b31-b52b-db33781f3dfa\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.438053 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.438066 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6674474f-8b80-4e0a-a6a0-9b894d2b6add-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.438080 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7hpf\" (UniqueName: \"kubernetes.io/projected/6674474f-8b80-4e0a-a6a0-9b894d2b6add-kube-api-access-s7hpf\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.438749 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8c06a2f0-3608-4b31-b52b-db33781f3dfa" (UID: "8c06a2f0-3608-4b31-b52b-db33781f3dfa"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.440107 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities" (OuterVolumeSpecName: "utilities") pod "d5545afd-0acb-4149-8e56-2804de448a28" (UID: "d5545afd-0acb-4149-8e56-2804de448a28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.444512 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr" (OuterVolumeSpecName: "kube-api-access-dgjzr") pod "d5545afd-0acb-4149-8e56-2804de448a28" (UID: "d5545afd-0acb-4149-8e56-2804de448a28"). InnerVolumeSpecName "kube-api-access-dgjzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.444792 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8c06a2f0-3608-4b31-b52b-db33781f3dfa" (UID: "8c06a2f0-3608-4b31-b52b-db33781f3dfa"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.445992 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd" (OuterVolumeSpecName: "kube-api-access-4lvsd") pod "8c06a2f0-3608-4b31-b52b-db33781f3dfa" (UID: "8c06a2f0-3608-4b31-b52b-db33781f3dfa"). InnerVolumeSpecName "kube-api-access-4lvsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.495856 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5545afd-0acb-4149-8e56-2804de448a28" (UID: "d5545afd-0acb-4149-8e56-2804de448a28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.511724 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.539848 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities\") pod \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.539919 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content\") pod \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.539950 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rtrf\" (UniqueName: \"kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf\") pod \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\" (UID: \"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540182 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540195 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lvsd\" (UniqueName: \"kubernetes.io/projected/8c06a2f0-3608-4b31-b52b-db33781f3dfa-kube-api-access-4lvsd\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540208 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5545afd-0acb-4149-8e56-2804de448a28-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540216 4921 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540226 4921 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c06a2f0-3608-4b31-b52b-db33781f3dfa-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540240 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgjzr\" (UniqueName: \"kubernetes.io/projected/d5545afd-0acb-4149-8e56-2804de448a28-kube-api-access-dgjzr\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.540686 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities" (OuterVolumeSpecName: "utilities") pod "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" (UID: "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.544853 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf" (OuterVolumeSpecName: "kube-api-access-8rtrf") pod "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" (UID: "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728"). InnerVolumeSpecName "kube-api-access-8rtrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.562468 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" (UID: "cb5cde82-e1a0-4d9f-800a-9aa70cc6e728"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641004 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhfs5\" (UniqueName: \"kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5\") pod \"6184e585-30f0-42d3-996f-ff66dba239c5\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641108 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities\") pod \"6184e585-30f0-42d3-996f-ff66dba239c5\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641168 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content\") pod \"6184e585-30f0-42d3-996f-ff66dba239c5\" (UID: \"6184e585-30f0-42d3-996f-ff66dba239c5\") " Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641492 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641513 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.641529 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rtrf\" (UniqueName: \"kubernetes.io/projected/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728-kube-api-access-8rtrf\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.643532 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5" (OuterVolumeSpecName: "kube-api-access-lhfs5") pod "6184e585-30f0-42d3-996f-ff66dba239c5" (UID: "6184e585-30f0-42d3-996f-ff66dba239c5"). InnerVolumeSpecName "kube-api-access-lhfs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.643857 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities" (OuterVolumeSpecName: "utilities") pod "6184e585-30f0-42d3-996f-ff66dba239c5" (UID: "6184e585-30f0-42d3-996f-ff66dba239c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.659233 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8s84l"] Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.742801 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.742991 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhfs5\" (UniqueName: \"kubernetes.io/projected/6184e585-30f0-42d3-996f-ff66dba239c5-kube-api-access-lhfs5\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.793022 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6184e585-30f0-42d3-996f-ff66dba239c5" (UID: "6184e585-30f0-42d3-996f-ff66dba239c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:25 crc kubenswrapper[4921]: I0103 03:47:25.844376 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6184e585-30f0-42d3-996f-ff66dba239c5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.084232 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qh9w2" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.084387 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qh9w2" event={"ID":"d5545afd-0acb-4149-8e56-2804de448a28","Type":"ContainerDied","Data":"85d2bc7e9317e1e4acc8f2cd96bcd1d3953612c03382ef48829d4c58b505acd6"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.084922 4921 scope.go:117] "RemoveContainer" containerID="9b06195fbbe2ff51f355597f1d1b8d19213b9a06f8c1fbaf4b7e0bda108e83a7" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.086179 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" event={"ID":"549a7ea1-10f1-4e90-a2e3-9ef968ce88da","Type":"ContainerStarted","Data":"35a710c9d8d0f5122e51e16bdbbf166aa2592b459e6c4f970b9004c563bbe54a"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.086226 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" event={"ID":"549a7ea1-10f1-4e90-a2e3-9ef968ce88da","Type":"ContainerStarted","Data":"ff9ab88468be50e1d26fecb4ed05c748c02a45865a246a55be8d058dd7a2762d"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.086464 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.087663 4921 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8s84l container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" start-of-body= Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.087712 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" podUID="549a7ea1-10f1-4e90-a2e3-9ef968ce88da" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.088949 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-khcdk" event={"ID":"cb5cde82-e1a0-4d9f-800a-9aa70cc6e728","Type":"ContainerDied","Data":"d110296cbd1a7409b1d96d72ed6446e215987f72708d7e920f1df145f4299dfe"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.088994 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-khcdk" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.091884 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z668n" event={"ID":"6184e585-30f0-42d3-996f-ff66dba239c5","Type":"ContainerDied","Data":"45fa296cde242a933df2734848ef92f7de5250cc394eddbf2e0d7279429032ec"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.091939 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z668n" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.096054 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.096199 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zng4m" event={"ID":"8c06a2f0-3608-4b31-b52b-db33781f3dfa","Type":"ContainerDied","Data":"8b1646238354ff8601733efe5ffc25147c15d1fae9edbec8c4b4b3218d4b45b6"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.106791 4921 scope.go:117] "RemoveContainer" containerID="0181ebb9968161926091d16d487a9a85cba9819b4402adfe5c2dcc305a0f0d5b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.108202 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj6h6" event={"ID":"6674474f-8b80-4e0a-a6a0-9b894d2b6add","Type":"ContainerDied","Data":"762e748c81c0f98162fa75d2c3241047a73b3185fb5fffee0464f9a5e112c4b4"} Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.108386 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj6h6" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.124184 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" podStartSLOduration=2.12415698 podStartE2EDuration="2.12415698s" podCreationTimestamp="2026-01-03 03:47:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:47:26.115403959 +0000 UTC m=+381.726830813" watchObservedRunningTime="2026-01-03 03:47:26.12415698 +0000 UTC m=+381.735583804" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.125379 4921 scope.go:117] "RemoveContainer" containerID="ed0905f013ed3acfd1db072ee0ca00197fbd85cbbbb08d88de6b53e218a92c9e" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.153117 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.165062 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z668n"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.166080 4921 scope.go:117] "RemoveContainer" containerID="e0a67b499b6b7f6b5fa3fe44f7fcee9aad934fd1fabc5777d1212f39035a3ea0" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.168887 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.180541 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qh9w2"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.186777 4921 scope.go:117] "RemoveContainer" containerID="3eabdcb5341d572a13d08f46410f97fd2a71fc6ea1da5153476a1d0e726f764f" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.189318 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.197015 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-khcdk"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.206758 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.211215 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zng4m"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.215106 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.217783 4921 scope.go:117] "RemoveContainer" containerID="f8c0afadc222a3b58cb5425ff75484f4b12618b6727f2d519356c173ccbd1d30" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.218385 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jj6h6"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.232157 4921 scope.go:117] "RemoveContainer" containerID="5d7c9ae78bdd293cdcc7cce232469d5a3a475c36d2d0c820ec432ef573c5154d" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.251136 4921 scope.go:117] "RemoveContainer" containerID="a1a7346f137c8744f39c26aa11957113707e346d1d5681f96c613bf2692ad8fa" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.270570 4921 scope.go:117] "RemoveContainer" containerID="6fcc382a23c32a92d08ae0b964139cd0cddfb9b28de2cf3287f7232dd631c3c3" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.288577 4921 scope.go:117] "RemoveContainer" containerID="e4b3743478d8831b4d8672f7e93d858e303d848e0f894b87dc8f48e5cfc04bf8" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.303109 4921 scope.go:117] "RemoveContainer" containerID="828eb13878805f6e8d371ef273a8950a0e7dce623d74d3801ce5528cc5574491" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.328818 4921 scope.go:117] "RemoveContainer" containerID="75c3d028166c131b673e9e268b1da5f87d168001a980eaf1e4be19a067c39d51" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.350249 4921 scope.go:117] "RemoveContainer" containerID="e534eaa8fa63e3d9bd982c4a5d07280711d3560dff64c0a0ac3b66c164b14533" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611232 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kkq2b"] Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611695 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611714 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611727 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611735 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611750 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611758 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611769 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611777 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611785 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611792 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611803 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611809 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611821 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerName="marketplace-operator" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611828 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerName="marketplace-operator" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611836 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611843 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="extract-utilities" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611851 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611861 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611873 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611881 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611889 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611897 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611910 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611918 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="extract-content" Jan 03 03:47:26 crc kubenswrapper[4921]: E0103 03:47:26.611928 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.611936 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.612062 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" containerName="marketplace-operator" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.612077 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5545afd-0acb-4149-8e56-2804de448a28" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.612090 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.612099 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.612109 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" containerName="registry-server" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.613203 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.617751 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.628682 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kkq2b"] Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.759054 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-utilities\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.759143 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-catalog-content\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.759224 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k88gb\" (UniqueName: \"kubernetes.io/projected/107e4592-8633-46d9-867c-c56028d7b8e8-kube-api-access-k88gb\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.861124 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-utilities\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.861193 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-catalog-content\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.861289 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k88gb\" (UniqueName: \"kubernetes.io/projected/107e4592-8633-46d9-867c-c56028d7b8e8-kube-api-access-k88gb\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.862757 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-utilities\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.863187 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/107e4592-8633-46d9-867c-c56028d7b8e8-catalog-content\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.889255 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6184e585-30f0-42d3-996f-ff66dba239c5" path="/var/lib/kubelet/pods/6184e585-30f0-42d3-996f-ff66dba239c5/volumes" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.890261 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6674474f-8b80-4e0a-a6a0-9b894d2b6add" path="/var/lib/kubelet/pods/6674474f-8b80-4e0a-a6a0-9b894d2b6add/volumes" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.890966 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c06a2f0-3608-4b31-b52b-db33781f3dfa" path="/var/lib/kubelet/pods/8c06a2f0-3608-4b31-b52b-db33781f3dfa/volumes" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.891898 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb5cde82-e1a0-4d9f-800a-9aa70cc6e728" path="/var/lib/kubelet/pods/cb5cde82-e1a0-4d9f-800a-9aa70cc6e728/volumes" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.892517 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5545afd-0acb-4149-8e56-2804de448a28" path="/var/lib/kubelet/pods/d5545afd-0acb-4149-8e56-2804de448a28/volumes" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.894168 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k88gb\" (UniqueName: \"kubernetes.io/projected/107e4592-8633-46d9-867c-c56028d7b8e8-kube-api-access-k88gb\") pod \"redhat-marketplace-kkq2b\" (UID: \"107e4592-8633-46d9-867c-c56028d7b8e8\") " pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:26 crc kubenswrapper[4921]: I0103 03:47:26.931794 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.126365 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8s84l" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.213861 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kxb2v"] Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.218468 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.219425 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxb2v"] Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.220472 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.370213 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-catalog-content\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.370384 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq8vd\" (UniqueName: \"kubernetes.io/projected/c05c8aa8-a90c-45ec-9d86-72760b49dd24-kube-api-access-pq8vd\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.370454 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-utilities\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.398676 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kkq2b"] Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.472173 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-catalog-content\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.472263 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq8vd\" (UniqueName: \"kubernetes.io/projected/c05c8aa8-a90c-45ec-9d86-72760b49dd24-kube-api-access-pq8vd\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.472319 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-utilities\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.472768 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-utilities\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.473038 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05c8aa8-a90c-45ec-9d86-72760b49dd24-catalog-content\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.493328 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq8vd\" (UniqueName: \"kubernetes.io/projected/c05c8aa8-a90c-45ec-9d86-72760b49dd24-kube-api-access-pq8vd\") pod \"certified-operators-kxb2v\" (UID: \"c05c8aa8-a90c-45ec-9d86-72760b49dd24\") " pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.535906 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:27 crc kubenswrapper[4921]: I0103 03:47:27.915523 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxb2v"] Jan 03 03:47:28 crc kubenswrapper[4921]: I0103 03:47:28.126804 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxb2v" event={"ID":"c05c8aa8-a90c-45ec-9d86-72760b49dd24","Type":"ContainerStarted","Data":"0166763b31a69a4afaf6b278791dc9164d9f24e88234089786d26dafda054743"} Jan 03 03:47:28 crc kubenswrapper[4921]: I0103 03:47:28.129174 4921 generic.go:334] "Generic (PLEG): container finished" podID="107e4592-8633-46d9-867c-c56028d7b8e8" containerID="a4c85ef82a8f9bdbd5542172e2e64585aee80c564c297da2b0b0fd48952f3a7c" exitCode=0 Jan 03 03:47:28 crc kubenswrapper[4921]: I0103 03:47:28.129554 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kkq2b" event={"ID":"107e4592-8633-46d9-867c-c56028d7b8e8","Type":"ContainerDied","Data":"a4c85ef82a8f9bdbd5542172e2e64585aee80c564c297da2b0b0fd48952f3a7c"} Jan 03 03:47:28 crc kubenswrapper[4921]: I0103 03:47:28.129607 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kkq2b" event={"ID":"107e4592-8633-46d9-867c-c56028d7b8e8","Type":"ContainerStarted","Data":"acbcf52fe8df7de5dd51b9995aaa6c788da5ea27cf1c802c1bda7548917e2476"} Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.009108 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5dqt"] Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.010856 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.013121 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.037834 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5dqt"] Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.094319 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-catalog-content\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.094446 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx7j9\" (UniqueName: \"kubernetes.io/projected/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-kube-api-access-xx7j9\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.094501 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-utilities\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.136805 4921 generic.go:334] "Generic (PLEG): container finished" podID="c05c8aa8-a90c-45ec-9d86-72760b49dd24" containerID="31b5b272d8e7561b5a38fa3fa471f55e8fb1848c5fedf7f0f246408866f97b5b" exitCode=0 Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.136879 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxb2v" event={"ID":"c05c8aa8-a90c-45ec-9d86-72760b49dd24","Type":"ContainerDied","Data":"31b5b272d8e7561b5a38fa3fa471f55e8fb1848c5fedf7f0f246408866f97b5b"} Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.195836 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-catalog-content\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.195923 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx7j9\" (UniqueName: \"kubernetes.io/projected/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-kube-api-access-xx7j9\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.195973 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-utilities\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.196649 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-catalog-content\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.196683 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-utilities\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.226787 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx7j9\" (UniqueName: \"kubernetes.io/projected/e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286-kube-api-access-xx7j9\") pod \"redhat-operators-r5dqt\" (UID: \"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286\") " pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.345419 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.614586 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bmtsf"] Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.617609 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.620979 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.622726 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bmtsf"] Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.711988 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tvzz\" (UniqueName: \"kubernetes.io/projected/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-kube-api-access-4tvzz\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.712532 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-utilities\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.712618 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-catalog-content\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.774994 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5dqt"] Jan 03 03:47:29 crc kubenswrapper[4921]: W0103 03:47:29.781983 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0c419ea_fdfa_4b5a_90ff_9bcbeb5a0286.slice/crio-486ecee35d1532eb14b8a90149f38716c4563f8d8ef91c3b561faffe1d9ebdd4 WatchSource:0}: Error finding container 486ecee35d1532eb14b8a90149f38716c4563f8d8ef91c3b561faffe1d9ebdd4: Status 404 returned error can't find the container with id 486ecee35d1532eb14b8a90149f38716c4563f8d8ef91c3b561faffe1d9ebdd4 Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.813559 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-catalog-content\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.813631 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tvzz\" (UniqueName: \"kubernetes.io/projected/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-kube-api-access-4tvzz\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.813669 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-utilities\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.814234 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-utilities\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.814313 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-catalog-content\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.836555 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tvzz\" (UniqueName: \"kubernetes.io/projected/0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4-kube-api-access-4tvzz\") pod \"community-operators-bmtsf\" (UID: \"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4\") " pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:29 crc kubenswrapper[4921]: I0103 03:47:29.938768 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.165395 4921 generic.go:334] "Generic (PLEG): container finished" podID="c05c8aa8-a90c-45ec-9d86-72760b49dd24" containerID="2c6355e96ef0849f2258b36103f1f050eab418b8aa302c944bd545cb572e9ee6" exitCode=0 Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.167500 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxb2v" event={"ID":"c05c8aa8-a90c-45ec-9d86-72760b49dd24","Type":"ContainerDied","Data":"2c6355e96ef0849f2258b36103f1f050eab418b8aa302c944bd545cb572e9ee6"} Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.172199 4921 generic.go:334] "Generic (PLEG): container finished" podID="107e4592-8633-46d9-867c-c56028d7b8e8" containerID="ec8be9278ef7c2266fa6474f5aa58f37af4055a583a405ce80e2a9188bf91815" exitCode=0 Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.172252 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kkq2b" event={"ID":"107e4592-8633-46d9-867c-c56028d7b8e8","Type":"ContainerDied","Data":"ec8be9278ef7c2266fa6474f5aa58f37af4055a583a405ce80e2a9188bf91815"} Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.178307 4921 generic.go:334] "Generic (PLEG): container finished" podID="e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286" containerID="ab47ad8aa45741f8f8b7a40846aaf1f74a819fa739ad194c51340a98167d22e4" exitCode=0 Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.178353 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5dqt" event={"ID":"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286","Type":"ContainerDied","Data":"ab47ad8aa45741f8f8b7a40846aaf1f74a819fa739ad194c51340a98167d22e4"} Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.178384 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5dqt" event={"ID":"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286","Type":"ContainerStarted","Data":"486ecee35d1532eb14b8a90149f38716c4563f8d8ef91c3b561faffe1d9ebdd4"} Jan 03 03:47:30 crc kubenswrapper[4921]: I0103 03:47:30.445682 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bmtsf"] Jan 03 03:47:30 crc kubenswrapper[4921]: W0103 03:47:30.454487 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c4a9ea4_8edd_42c3_af24_d96bb6d8e4e4.slice/crio-435d8e4ead0a35231b4053f1de3649dbb2b6781359807408db9626eaa9cffbdb WatchSource:0}: Error finding container 435d8e4ead0a35231b4053f1de3649dbb2b6781359807408db9626eaa9cffbdb: Status 404 returned error can't find the container with id 435d8e4ead0a35231b4053f1de3649dbb2b6781359807408db9626eaa9cffbdb Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.175715 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.176126 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.187489 4921 generic.go:334] "Generic (PLEG): container finished" podID="0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4" containerID="4c1bb38761ec765641217958eb31e81c85af93a55a5bde24cfd733c58daa2e70" exitCode=0 Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.187535 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmtsf" event={"ID":"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4","Type":"ContainerDied","Data":"4c1bb38761ec765641217958eb31e81c85af93a55a5bde24cfd733c58daa2e70"} Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.187604 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmtsf" event={"ID":"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4","Type":"ContainerStarted","Data":"435d8e4ead0a35231b4053f1de3649dbb2b6781359807408db9626eaa9cffbdb"} Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.191736 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxb2v" event={"ID":"c05c8aa8-a90c-45ec-9d86-72760b49dd24","Type":"ContainerStarted","Data":"972f1a602f7b4c1dbd393543c04248a0f082fbd985e232927979048c28ed649a"} Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.194098 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kkq2b" event={"ID":"107e4592-8633-46d9-867c-c56028d7b8e8","Type":"ContainerStarted","Data":"bab4ac653d89ab718212853414b8f79b06fe19f1a78d2fe7501252a61126e8ee"} Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.196687 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5dqt" event={"ID":"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286","Type":"ContainerStarted","Data":"7d6adc0fc35bf841869a35e1ec066886d16d72f3e0b0ce81788f3d166d7e4bf8"} Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.259292 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kkq2b" podStartSLOduration=2.471086743 podStartE2EDuration="5.259259733s" podCreationTimestamp="2026-01-03 03:47:26 +0000 UTC" firstStartedPulling="2026-01-03 03:47:28.130987397 +0000 UTC m=+383.742414221" lastFinishedPulling="2026-01-03 03:47:30.919160367 +0000 UTC m=+386.530587211" observedRunningTime="2026-01-03 03:47:31.256558039 +0000 UTC m=+386.867984863" watchObservedRunningTime="2026-01-03 03:47:31.259259733 +0000 UTC m=+386.870686557" Jan 03 03:47:31 crc kubenswrapper[4921]: I0103 03:47:31.280404 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kxb2v" podStartSLOduration=2.649835512 podStartE2EDuration="4.280384666s" podCreationTimestamp="2026-01-03 03:47:27 +0000 UTC" firstStartedPulling="2026-01-03 03:47:29.139292256 +0000 UTC m=+384.750719080" lastFinishedPulling="2026-01-03 03:47:30.76984141 +0000 UTC m=+386.381268234" observedRunningTime="2026-01-03 03:47:31.27835226 +0000 UTC m=+386.889779094" watchObservedRunningTime="2026-01-03 03:47:31.280384666 +0000 UTC m=+386.891811490" Jan 03 03:47:32 crc kubenswrapper[4921]: I0103 03:47:32.205370 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmtsf" event={"ID":"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4","Type":"ContainerStarted","Data":"23aa56e98575620cdd6b0ee4e5629cf812d3cd13b148c7a893d4d6e6d59d9b38"} Jan 03 03:47:32 crc kubenswrapper[4921]: I0103 03:47:32.207147 4921 generic.go:334] "Generic (PLEG): container finished" podID="e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286" containerID="7d6adc0fc35bf841869a35e1ec066886d16d72f3e0b0ce81788f3d166d7e4bf8" exitCode=0 Jan 03 03:47:32 crc kubenswrapper[4921]: I0103 03:47:32.207228 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5dqt" event={"ID":"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286","Type":"ContainerDied","Data":"7d6adc0fc35bf841869a35e1ec066886d16d72f3e0b0ce81788f3d166d7e4bf8"} Jan 03 03:47:33 crc kubenswrapper[4921]: I0103 03:47:33.214764 4921 generic.go:334] "Generic (PLEG): container finished" podID="0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4" containerID="23aa56e98575620cdd6b0ee4e5629cf812d3cd13b148c7a893d4d6e6d59d9b38" exitCode=0 Jan 03 03:47:33 crc kubenswrapper[4921]: I0103 03:47:33.214823 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmtsf" event={"ID":"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4","Type":"ContainerDied","Data":"23aa56e98575620cdd6b0ee4e5629cf812d3cd13b148c7a893d4d6e6d59d9b38"} Jan 03 03:47:33 crc kubenswrapper[4921]: I0103 03:47:33.220881 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5dqt" event={"ID":"e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286","Type":"ContainerStarted","Data":"38bd03fb8e0b5c9e2b27fd4666f2f8126bdc48ab00fde45002a09838a92c12a3"} Jan 03 03:47:33 crc kubenswrapper[4921]: I0103 03:47:33.252710 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5dqt" podStartSLOduration=2.6988373340000003 podStartE2EDuration="5.252685722s" podCreationTimestamp="2026-01-03 03:47:28 +0000 UTC" firstStartedPulling="2026-01-03 03:47:30.18003845 +0000 UTC m=+385.791465274" lastFinishedPulling="2026-01-03 03:47:32.733886848 +0000 UTC m=+388.345313662" observedRunningTime="2026-01-03 03:47:33.249984147 +0000 UTC m=+388.861410971" watchObservedRunningTime="2026-01-03 03:47:33.252685722 +0000 UTC m=+388.864112546" Jan 03 03:47:34 crc kubenswrapper[4921]: I0103 03:47:34.230210 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bmtsf" event={"ID":"0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4","Type":"ContainerStarted","Data":"d27c571616e91221c18160b977bdd4d5b2e62a0424d3b64b00e99b84060ce0bd"} Jan 03 03:47:34 crc kubenswrapper[4921]: I0103 03:47:34.253264 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bmtsf" podStartSLOduration=2.806472489 podStartE2EDuration="5.253238426s" podCreationTimestamp="2026-01-03 03:47:29 +0000 UTC" firstStartedPulling="2026-01-03 03:47:31.189751127 +0000 UTC m=+386.801177951" lastFinishedPulling="2026-01-03 03:47:33.636517064 +0000 UTC m=+389.247943888" observedRunningTime="2026-01-03 03:47:34.251051266 +0000 UTC m=+389.862478090" watchObservedRunningTime="2026-01-03 03:47:34.253238426 +0000 UTC m=+389.864665250" Jan 03 03:47:36 crc kubenswrapper[4921]: I0103 03:47:36.932757 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:36 crc kubenswrapper[4921]: I0103 03:47:36.933327 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:36 crc kubenswrapper[4921]: I0103 03:47:36.986258 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:37 crc kubenswrapper[4921]: I0103 03:47:37.298664 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kkq2b" Jan 03 03:47:37 crc kubenswrapper[4921]: I0103 03:47:37.536609 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:37 crc kubenswrapper[4921]: I0103 03:47:37.536678 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:37 crc kubenswrapper[4921]: I0103 03:47:37.585437 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:38 crc kubenswrapper[4921]: I0103 03:47:38.299393 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kxb2v" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.346414 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.347381 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.401039 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.939880 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.940432 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:39 crc kubenswrapper[4921]: I0103 03:47:39.990233 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:40 crc kubenswrapper[4921]: I0103 03:47:40.305230 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5dqt" Jan 03 03:47:40 crc kubenswrapper[4921]: I0103 03:47:40.312753 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bmtsf" Jan 03 03:47:46 crc kubenswrapper[4921]: I0103 03:47:46.324957 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" podUID="c7d5610b-f8e0-4984-9907-ce33eb526161" containerName="registry" containerID="cri-o://9917ab2fceb8f7c406980ab44ac01e03f9068039069bcc4c71ab0e6ab6a99dfc" gracePeriod=30 Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.306744 4921 generic.go:334] "Generic (PLEG): container finished" podID="c7d5610b-f8e0-4984-9907-ce33eb526161" containerID="9917ab2fceb8f7c406980ab44ac01e03f9068039069bcc4c71ab0e6ab6a99dfc" exitCode=0 Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.306829 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" event={"ID":"c7d5610b-f8e0-4984-9907-ce33eb526161","Type":"ContainerDied","Data":"9917ab2fceb8f7c406980ab44ac01e03f9068039069bcc4c71ab0e6ab6a99dfc"} Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.359925 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.486504 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.486660 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr6cz\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.486763 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.486832 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.487611 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.487220 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.487902 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.488151 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.488222 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls\") pod \"c7d5610b-f8e0-4984-9907-ce33eb526161\" (UID: \"c7d5610b-f8e0-4984-9907-ce33eb526161\") " Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.488624 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.488739 4921 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.488754 4921 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.494680 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.495122 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz" (OuterVolumeSpecName: "kube-api-access-mr6cz") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "kube-api-access-mr6cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.496160 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.499112 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.505536 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.510264 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c7d5610b-f8e0-4984-9907-ce33eb526161" (UID: "c7d5610b-f8e0-4984-9907-ce33eb526161"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.590051 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr6cz\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-kube-api-access-mr6cz\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.590087 4921 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c7d5610b-f8e0-4984-9907-ce33eb526161-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.590096 4921 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.590104 4921 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c7d5610b-f8e0-4984-9907-ce33eb526161-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:47 crc kubenswrapper[4921]: I0103 03:47:47.590115 4921 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c7d5610b-f8e0-4984-9907-ce33eb526161-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.316066 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" event={"ID":"c7d5610b-f8e0-4984-9907-ce33eb526161","Type":"ContainerDied","Data":"ea2800b0811a233029d5818631cee1b34a29f925ff5360aee493a3236a6150a1"} Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.316136 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-mvvcd" Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.316144 4921 scope.go:117] "RemoveContainer" containerID="9917ab2fceb8f7c406980ab44ac01e03f9068039069bcc4c71ab0e6ab6a99dfc" Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.353162 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.357463 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-mvvcd"] Jan 03 03:47:48 crc kubenswrapper[4921]: I0103 03:47:48.895709 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d5610b-f8e0-4984-9907-ce33eb526161" path="/var/lib/kubelet/pods/c7d5610b-f8e0-4984-9907-ce33eb526161/volumes" Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.176147 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.176616 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.176687 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.177662 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.177984 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329" gracePeriod=600 Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.407803 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329" exitCode=0 Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.407870 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329"} Jan 03 03:48:01 crc kubenswrapper[4921]: I0103 03:48:01.407928 4921 scope.go:117] "RemoveContainer" containerID="638dfe07911c70ff91a65878c0a09f6506945f534e82b5abc501a27be2a94625" Jan 03 03:48:02 crc kubenswrapper[4921]: I0103 03:48:02.419591 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893"} Jan 03 03:50:01 crc kubenswrapper[4921]: I0103 03:50:01.176911 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:50:01 crc kubenswrapper[4921]: I0103 03:50:01.177926 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:50:31 crc kubenswrapper[4921]: I0103 03:50:31.176567 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:50:31 crc kubenswrapper[4921]: I0103 03:50:31.177336 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.176203 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.177327 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.177420 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.178479 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.178567 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893" gracePeriod=600 Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.655982 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893" exitCode=0 Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.656121 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893"} Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.656649 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2"} Jan 03 03:51:01 crc kubenswrapper[4921]: I0103 03:51:01.656692 4921 scope.go:117] "RemoveContainer" containerID="a9e031625dcc37affc12e48e49bf539d1c8b4f1ada146de8f8e545523c72d329" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.550831 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbb7k"] Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.551818 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-controller" containerID="cri-o://1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.551843 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="sbdb" containerID="cri-o://dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.551887 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.551985 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="nbdb" containerID="cri-o://4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.552014 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="northd" containerID="cri-o://fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.551997 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-acl-logging" containerID="cri-o://c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.552026 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-node" containerID="cri-o://2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.607449 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" containerID="cri-o://d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" gracePeriod=30 Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.853886 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/3.log" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.856780 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovn-acl-logging/0.log" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.857440 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovn-controller/0.log" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.858236 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.915676 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w52bw"] Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.915920 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-ovn-metrics" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.915937 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-ovn-metrics" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.915948 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-node" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.915957 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-node" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.915969 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.915978 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.915988 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.915996 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.916006 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kubecfg-setup" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.916017 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kubecfg-setup" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.917772 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.917853 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.917879 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.917933 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.917961 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="northd" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.917975 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="northd" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.918041 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918057 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.918122 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d5610b-f8e0-4984-9907-ce33eb526161" containerName="registry" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918143 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d5610b-f8e0-4984-9907-ce33eb526161" containerName="registry" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.918172 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="nbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918221 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="nbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.918321 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="sbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918346 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="sbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.918430 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-acl-logging" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918469 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-acl-logging" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918862 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-acl-logging" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918894 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-node" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918913 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918932 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="nbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918955 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="sbdb" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918975 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.918991 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d5610b-f8e0-4984-9907-ce33eb526161" containerName="registry" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919010 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovn-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919032 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="kube-rbac-proxy-ovn-metrics" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919052 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919067 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="northd" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919083 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919101 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: E0103 03:52:55.919320 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.919337 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerName="ovnkube-controller" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.923087 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994382 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994419 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994441 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994463 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994486 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994527 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994527 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994633 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgv5c\" (UniqueName: \"kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994651 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994667 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994683 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994700 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994724 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994751 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994768 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994790 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994821 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994838 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994854 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994869 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994882 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994914 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn\") pod \"bf34943c-bfe4-4411-af8a-189b14e35a82\" (UID: \"bf34943c-bfe4-4411-af8a-189b14e35a82\") " Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995031 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-bin\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995062 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995084 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-systemd-units\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995099 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-kubelet\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995129 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-env-overrides\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995143 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-netd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995158 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vnjc\" (UniqueName: \"kubernetes.io/projected/368bd5f7-ae1a-4333-857b-383f44500621-kube-api-access-7vnjc\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995176 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/368bd5f7-ae1a-4333-857b-383f44500621-ovn-node-metrics-cert\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995219 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-var-lib-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995246 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-config\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995260 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-log-socket\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995294 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-systemd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995313 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-netns\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995333 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-script-lib\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995350 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-etc-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995375 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995394 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-slash\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995420 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995439 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-ovn\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995464 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-node-log\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995492 4921 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995503 4921 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994678 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994921 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994945 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.994999 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995836 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995868 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995891 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.995912 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log" (OuterVolumeSpecName: "node-log") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996095 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996121 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996166 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket" (OuterVolumeSpecName: "log-socket") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996142 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996211 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash" (OuterVolumeSpecName: "host-slash") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996353 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:55 crc kubenswrapper[4921]: I0103 03:52:55.996781 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.013705 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.013820 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c" (OuterVolumeSpecName: "kube-api-access-mgv5c") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "kube-api-access-mgv5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.013863 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bf34943c-bfe4-4411-af8a-189b14e35a82" (UID: "bf34943c-bfe4-4411-af8a-189b14e35a82"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103158 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-node-log\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103239 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-bin\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103310 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103348 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-systemd-units\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103402 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-systemd-units\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103410 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103354 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-bin\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103355 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-node-log\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103411 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-kubelet\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103497 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-kubelet\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103610 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-env-overrides\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.103647 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-netd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105407 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vnjc\" (UniqueName: \"kubernetes.io/projected/368bd5f7-ae1a-4333-857b-383f44500621-kube-api-access-7vnjc\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105498 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/368bd5f7-ae1a-4333-857b-383f44500621-ovn-node-metrics-cert\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105574 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-var-lib-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105613 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-env-overrides\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105683 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-config\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105736 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-cni-netd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.105804 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-var-lib-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107081 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-log-socket\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107187 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-log-socket\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107241 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-systemd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107358 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-netns\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107403 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-systemd\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107429 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-script-lib\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107451 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-netns\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107484 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-etc-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107480 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-config\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107530 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-etc-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107544 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107587 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-slash\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107650 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-openvswitch\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107659 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107714 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-ovn\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107768 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-slash\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107829 4921 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107840 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-run-ovn\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107864 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/368bd5f7-ae1a-4333-857b-383f44500621-host-run-ovn-kubernetes\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107892 4921 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107912 4921 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107934 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgv5c\" (UniqueName: \"kubernetes.io/projected/bf34943c-bfe4-4411-af8a-189b14e35a82-kube-api-access-mgv5c\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107953 4921 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bf34943c-bfe4-4411-af8a-189b14e35a82-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107970 4921 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.107988 4921 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bf34943c-bfe4-4411-af8a-189b14e35a82-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108036 4921 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108058 4921 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108075 4921 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108088 4921 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108100 4921 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-log-socket\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108114 4921 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108126 4921 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-slash\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108140 4921 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108154 4921 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108166 4921 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-node-log\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108180 4921 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bf34943c-bfe4-4411-af8a-189b14e35a82-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.108855 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/368bd5f7-ae1a-4333-857b-383f44500621-ovnkube-script-lib\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.113375 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/368bd5f7-ae1a-4333-857b-383f44500621-ovn-node-metrics-cert\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.134866 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vnjc\" (UniqueName: \"kubernetes.io/projected/368bd5f7-ae1a-4333-857b-383f44500621-kube-api-access-7vnjc\") pod \"ovnkube-node-w52bw\" (UID: \"368bd5f7-ae1a-4333-857b-383f44500621\") " pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.237711 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.597159 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovnkube-controller/3.log" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.602011 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovn-acl-logging/0.log" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.603456 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbb7k_bf34943c-bfe4-4411-af8a-189b14e35a82/ovn-controller/0.log" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604304 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604340 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604351 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604361 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604369 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604377 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604394 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" exitCode=143 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604403 4921 generic.go:334] "Generic (PLEG): container finished" podID="bf34943c-bfe4-4411-af8a-189b14e35a82" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" exitCode=143 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604421 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604539 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604568 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604586 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604603 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604626 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604631 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604645 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604667 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604679 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604689 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604699 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604709 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604719 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604729 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604738 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604752 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604773 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604788 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604797 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604807 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604817 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604826 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604835 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604845 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604854 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604863 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604876 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604892 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604907 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604920 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604933 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604946 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604957 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604967 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.604990 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605003 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605013 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605028 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" event={"ID":"bf34943c-bfe4-4411-af8a-189b14e35a82","Type":"ContainerDied","Data":"59afa832eb9a37573ea31b7f77d861097a8dd7a8d8b4d635afa837e88844cc37"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605044 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605059 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605069 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605079 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605088 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605099 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605108 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605118 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605127 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.605137 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.607534 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbb7k" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.609166 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/2.log" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.609899 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/1.log" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.609932 4921 generic.go:334] "Generic (PLEG): container finished" podID="116d8b0a-baa0-4087-9a34-8e890ab8f8dc" containerID="956614e9298ecf848c23f3839a090463fa6a357c9bf13d8ba23ff555363d03cf" exitCode=2 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.609985 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerDied","Data":"956614e9298ecf848c23f3839a090463fa6a357c9bf13d8ba23ff555363d03cf"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.610022 4921 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.610722 4921 scope.go:117] "RemoveContainer" containerID="956614e9298ecf848c23f3839a090463fa6a357c9bf13d8ba23ff555363d03cf" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.610990 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-666ct_openshift-multus(116d8b0a-baa0-4087-9a34-8e890ab8f8dc)\"" pod="openshift-multus/multus-666ct" podUID="116d8b0a-baa0-4087-9a34-8e890ab8f8dc" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.614418 4921 generic.go:334] "Generic (PLEG): container finished" podID="368bd5f7-ae1a-4333-857b-383f44500621" containerID="f9cb66a349507f40096d2d30fd3a96d1567622bc886f52645063fdbccb644531" exitCode=0 Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.614452 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerDied","Data":"f9cb66a349507f40096d2d30fd3a96d1567622bc886f52645063fdbccb644531"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.614480 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"ad276c8eca8e391db279ce554c3e120748b30cb54932d0a5f62d28e46abc101c"} Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.638871 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.670556 4921 scope.go:117] "RemoveContainer" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.718565 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbb7k"] Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.720950 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbb7k"] Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.737552 4921 scope.go:117] "RemoveContainer" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.758967 4921 scope.go:117] "RemoveContainer" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.777249 4921 scope.go:117] "RemoveContainer" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.793249 4921 scope.go:117] "RemoveContainer" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.806539 4921 scope.go:117] "RemoveContainer" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.822446 4921 scope.go:117] "RemoveContainer" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.845176 4921 scope.go:117] "RemoveContainer" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.873074 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.873642 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.873676 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} err="failed to get container status \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.873700 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.873895 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": container with ID starting with 294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a not found: ID does not exist" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.873912 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} err="failed to get container status \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": rpc error: code = NotFound desc = could not find container \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": container with ID starting with 294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.873926 4921 scope.go:117] "RemoveContainer" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.874104 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": container with ID starting with dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005 not found: ID does not exist" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874120 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} err="failed to get container status \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": rpc error: code = NotFound desc = could not find container \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": container with ID starting with dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874133 4921 scope.go:117] "RemoveContainer" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.874345 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": container with ID starting with 4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45 not found: ID does not exist" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874361 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} err="failed to get container status \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": rpc error: code = NotFound desc = could not find container \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": container with ID starting with 4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874373 4921 scope.go:117] "RemoveContainer" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.874714 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": container with ID starting with fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf not found: ID does not exist" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874728 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} err="failed to get container status \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": rpc error: code = NotFound desc = could not find container \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": container with ID starting with fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874740 4921 scope.go:117] "RemoveContainer" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.874948 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": container with ID starting with 528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231 not found: ID does not exist" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874962 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} err="failed to get container status \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": rpc error: code = NotFound desc = could not find container \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": container with ID starting with 528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.874972 4921 scope.go:117] "RemoveContainer" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.875156 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": container with ID starting with 2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155 not found: ID does not exist" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.875169 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} err="failed to get container status \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": rpc error: code = NotFound desc = could not find container \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": container with ID starting with 2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.875181 4921 scope.go:117] "RemoveContainer" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.875606 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": container with ID starting with c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590 not found: ID does not exist" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.875623 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} err="failed to get container status \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": rpc error: code = NotFound desc = could not find container \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": container with ID starting with c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.875634 4921 scope.go:117] "RemoveContainer" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.876048 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": container with ID starting with 1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e not found: ID does not exist" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.876063 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} err="failed to get container status \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": rpc error: code = NotFound desc = could not find container \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": container with ID starting with 1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.876074 4921 scope.go:117] "RemoveContainer" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: E0103 03:52:56.876374 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": container with ID starting with 8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615 not found: ID does not exist" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.877952 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} err="failed to get container status \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": rpc error: code = NotFound desc = could not find container \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": container with ID starting with 8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.878027 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.878454 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} err="failed to get container status \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.878536 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.878868 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} err="failed to get container status \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": rpc error: code = NotFound desc = could not find container \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": container with ID starting with 294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.878948 4921 scope.go:117] "RemoveContainer" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.879294 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} err="failed to get container status \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": rpc error: code = NotFound desc = could not find container \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": container with ID starting with dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.879383 4921 scope.go:117] "RemoveContainer" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.879687 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} err="failed to get container status \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": rpc error: code = NotFound desc = could not find container \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": container with ID starting with 4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.879789 4921 scope.go:117] "RemoveContainer" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880073 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} err="failed to get container status \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": rpc error: code = NotFound desc = could not find container \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": container with ID starting with fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880097 4921 scope.go:117] "RemoveContainer" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880381 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} err="failed to get container status \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": rpc error: code = NotFound desc = could not find container \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": container with ID starting with 528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880461 4921 scope.go:117] "RemoveContainer" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880711 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} err="failed to get container status \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": rpc error: code = NotFound desc = could not find container \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": container with ID starting with 2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.880784 4921 scope.go:117] "RemoveContainer" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881063 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} err="failed to get container status \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": rpc error: code = NotFound desc = could not find container \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": container with ID starting with c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881136 4921 scope.go:117] "RemoveContainer" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881407 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} err="failed to get container status \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": rpc error: code = NotFound desc = could not find container \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": container with ID starting with 1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881486 4921 scope.go:117] "RemoveContainer" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881782 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} err="failed to get container status \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": rpc error: code = NotFound desc = could not find container \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": container with ID starting with 8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.881865 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.882120 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} err="failed to get container status \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.882189 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.882482 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} err="failed to get container status \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": rpc error: code = NotFound desc = could not find container \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": container with ID starting with 294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.882565 4921 scope.go:117] "RemoveContainer" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883007 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} err="failed to get container status \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": rpc error: code = NotFound desc = could not find container \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": container with ID starting with dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883091 4921 scope.go:117] "RemoveContainer" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883433 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} err="failed to get container status \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": rpc error: code = NotFound desc = could not find container \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": container with ID starting with 4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883456 4921 scope.go:117] "RemoveContainer" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883795 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} err="failed to get container status \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": rpc error: code = NotFound desc = could not find container \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": container with ID starting with fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.883888 4921 scope.go:117] "RemoveContainer" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.884185 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} err="failed to get container status \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": rpc error: code = NotFound desc = could not find container \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": container with ID starting with 528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.884300 4921 scope.go:117] "RemoveContainer" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.884637 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} err="failed to get container status \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": rpc error: code = NotFound desc = could not find container \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": container with ID starting with 2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.884728 4921 scope.go:117] "RemoveContainer" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885113 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} err="failed to get container status \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": rpc error: code = NotFound desc = could not find container \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": container with ID starting with c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885138 4921 scope.go:117] "RemoveContainer" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885437 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} err="failed to get container status \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": rpc error: code = NotFound desc = could not find container \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": container with ID starting with 1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885463 4921 scope.go:117] "RemoveContainer" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885776 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} err="failed to get container status \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": rpc error: code = NotFound desc = could not find container \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": container with ID starting with 8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.885800 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886040 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} err="failed to get container status \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886112 4921 scope.go:117] "RemoveContainer" containerID="294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886555 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a"} err="failed to get container status \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": rpc error: code = NotFound desc = could not find container \"294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a\": container with ID starting with 294a2cd4451df788f380a82d7551d584761befba6fba165f8caea5792026153a not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886579 4921 scope.go:117] "RemoveContainer" containerID="dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886858 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005"} err="failed to get container status \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": rpc error: code = NotFound desc = could not find container \"dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005\": container with ID starting with dbbd012862469478aac7a028405568eb03c65f72dadcd1fc0012f28a8a1ce005 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.886882 4921 scope.go:117] "RemoveContainer" containerID="4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887154 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45"} err="failed to get container status \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": rpc error: code = NotFound desc = could not find container \"4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45\": container with ID starting with 4d254fe2bde2b78028a3e0f1543278ff8e9a21ac4c42a1ebdbb8bb84a5ebbc45 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887176 4921 scope.go:117] "RemoveContainer" containerID="fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887403 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf"} err="failed to get container status \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": rpc error: code = NotFound desc = could not find container \"fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf\": container with ID starting with fb644cdd8782c6313666b8c34a35776680e5f3baa1dba59cc66e09ff0c82eddf not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887508 4921 scope.go:117] "RemoveContainer" containerID="528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887840 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231"} err="failed to get container status \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": rpc error: code = NotFound desc = could not find container \"528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231\": container with ID starting with 528ea1a96d689d9784e060ded192085dbc92224725bd1db7f8718e756f2e2231 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.887920 4921 scope.go:117] "RemoveContainer" containerID="2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.888261 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155"} err="failed to get container status \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": rpc error: code = NotFound desc = could not find container \"2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155\": container with ID starting with 2c42e5ef49508ac695dc10469a7bfaecebb833c942dec6f4293fa5868ec45155 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.888370 4921 scope.go:117] "RemoveContainer" containerID="c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.888630 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590"} err="failed to get container status \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": rpc error: code = NotFound desc = could not find container \"c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590\": container with ID starting with c4e4e3ab03b73ebd1665388e09931d7e7e07138145381a29bb6e0e6faf92a590 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.888711 4921 scope.go:117] "RemoveContainer" containerID="1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.889091 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e"} err="failed to get container status \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": rpc error: code = NotFound desc = could not find container \"1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e\": container with ID starting with 1dceb3665f248275f50e031839bbcf23a4dba056e04ce7481909cb8eaf7a419e not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.889186 4921 scope.go:117] "RemoveContainer" containerID="8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.889520 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615"} err="failed to get container status \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": rpc error: code = NotFound desc = could not find container \"8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615\": container with ID starting with 8b1b05d2a88092e3a7a2826e151cd1d0a27a6b008a50b99def20bdd51d847615 not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.889546 4921 scope.go:117] "RemoveContainer" containerID="d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.889835 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df"} err="failed to get container status \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": rpc error: code = NotFound desc = could not find container \"d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df\": container with ID starting with d47dead88a40ee3d81abc58c246fe4df7318939797e34ec761b8f17e1cb309df not found: ID does not exist" Jan 03 03:52:56 crc kubenswrapper[4921]: I0103 03:52:56.908970 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf34943c-bfe4-4411-af8a-189b14e35a82" path="/var/lib/kubelet/pods/bf34943c-bfe4-4411-af8a-189b14e35a82/volumes" Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625127 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"7dde9da7a9be504067c1a5df2dd7e0cf25cf01d8750de9537432cd29dbebf0ec"} Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625668 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"51445cf39be67d0c2ecc50dd7fcdfb8553bccfb77ce0ab6e32a6571c5b0f8ca6"} Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625688 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"429a5cb85099c75e3529990171242d9f33aaf1880042c9b7203cad0cd004c2b2"} Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625717 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"2be6d4186dac37a67444e551641a0cdb72ac02f834cc35c0e27a22e2348b7c4a"} Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625730 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"5fecadae23e1410f00fe62f4548d55130df9a2d7cdca34c902cbfee600c80c2d"} Jan 03 03:52:57 crc kubenswrapper[4921]: I0103 03:52:57.625741 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"e2d308431f040ea6e406ec899c971fbe4a96afb1a751baf4c1c59e44b375357c"} Jan 03 03:53:00 crc kubenswrapper[4921]: I0103 03:53:00.651726 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"04edae25957bd7a484cc831bb31ffe1609243b6fd59d66124df68993f74cd05c"} Jan 03 03:53:01 crc kubenswrapper[4921]: I0103 03:53:01.176439 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:53:01 crc kubenswrapper[4921]: I0103 03:53:01.176782 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.668516 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" event={"ID":"368bd5f7-ae1a-4333-857b-383f44500621","Type":"ContainerStarted","Data":"b59cef287259da15ae533f7f4cff255d7db5f528abd91ab998ce00f4d0953710"} Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.669562 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.669597 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.669608 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.703924 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.706517 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:02 crc kubenswrapper[4921]: I0103 03:53:02.707068 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" podStartSLOduration=7.707039858 podStartE2EDuration="7.707039858s" podCreationTimestamp="2026-01-03 03:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:53:02.70495519 +0000 UTC m=+718.316382054" watchObservedRunningTime="2026-01-03 03:53:02.707039858 +0000 UTC m=+718.318466702" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.183524 4921 scope.go:117] "RemoveContainer" containerID="5d74740e054f3062e8e2f6b695467a9ced26c1d2cb8f856be4491feb985b0b0b" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.188140 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4"] Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.189652 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.193991 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.199884 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4"] Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.237540 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.237639 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.237678 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpdg\" (UniqueName: \"kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.338953 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.339008 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.339039 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpdg\" (UniqueName: \"kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.339895 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.340972 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.365752 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpdg\" (UniqueName: \"kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.543292 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.579938 4921 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(516983781d4401599e1a1a034a756c4ca33fb32f13fc8fbed94698361c52ffca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.580398 4921 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(516983781d4401599e1a1a034a756c4ca33fb32f13fc8fbed94698361c52ffca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.580441 4921 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(516983781d4401599e1a1a034a756c4ca33fb32f13fc8fbed94698361c52ffca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.580517 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace(da345cb0-e3ff-4397-bc44-1e0e4b419c59)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace(da345cb0-e3ff-4397-bc44-1e0e4b419c59)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(516983781d4401599e1a1a034a756c4ca33fb32f13fc8fbed94698361c52ffca): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.687904 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/2.log" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.688080 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: I0103 03:53:05.688913 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.739488 4921 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(b1018841ee55d55d6798826108f7e4a45231b6f676d93953f5b68c1ae204c021): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.739567 4921 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(b1018841ee55d55d6798826108f7e4a45231b6f676d93953f5b68c1ae204c021): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.739595 4921 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(b1018841ee55d55d6798826108f7e4a45231b6f676d93953f5b68c1ae204c021): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:05 crc kubenswrapper[4921]: E0103 03:53:05.739652 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace(da345cb0-e3ff-4397-bc44-1e0e4b419c59)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace(da345cb0-e3ff-4397-bc44-1e0e4b419c59)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_openshift-marketplace_da345cb0-e3ff-4397-bc44-1e0e4b419c59_0(b1018841ee55d55d6798826108f7e4a45231b6f676d93953f5b68c1ae204c021): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" Jan 03 03:53:11 crc kubenswrapper[4921]: I0103 03:53:11.883859 4921 scope.go:117] "RemoveContainer" containerID="956614e9298ecf848c23f3839a090463fa6a357c9bf13d8ba23ff555363d03cf" Jan 03 03:53:13 crc kubenswrapper[4921]: I0103 03:53:13.748070 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-666ct_116d8b0a-baa0-4087-9a34-8e890ab8f8dc/kube-multus/2.log" Jan 03 03:53:13 crc kubenswrapper[4921]: I0103 03:53:13.748839 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-666ct" event={"ID":"116d8b0a-baa0-4087-9a34-8e890ab8f8dc","Type":"ContainerStarted","Data":"7242dad555b4d0cdd6d3abbf58e92e82a1b1ea22ef5f640a5056dfa241689bbf"} Jan 03 03:53:17 crc kubenswrapper[4921]: I0103 03:53:17.883689 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:17 crc kubenswrapper[4921]: I0103 03:53:17.885239 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:18 crc kubenswrapper[4921]: I0103 03:53:18.089407 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4"] Jan 03 03:53:18 crc kubenswrapper[4921]: I0103 03:53:18.781926 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerStarted","Data":"23047fcd3d2db4f4a4d58b052fa42280ae63d0b51da10c53693daf24c69616a0"} Jan 03 03:53:18 crc kubenswrapper[4921]: I0103 03:53:18.782459 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerStarted","Data":"9e2a8234dc652329962cbc29374dadbcb531feb466b34638067b149ce39f2a0e"} Jan 03 03:53:19 crc kubenswrapper[4921]: I0103 03:53:19.789779 4921 generic.go:334] "Generic (PLEG): container finished" podID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerID="23047fcd3d2db4f4a4d58b052fa42280ae63d0b51da10c53693daf24c69616a0" exitCode=0 Jan 03 03:53:19 crc kubenswrapper[4921]: I0103 03:53:19.789844 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerDied","Data":"23047fcd3d2db4f4a4d58b052fa42280ae63d0b51da10c53693daf24c69616a0"} Jan 03 03:53:19 crc kubenswrapper[4921]: I0103 03:53:19.791949 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 03:53:21 crc kubenswrapper[4921]: I0103 03:53:21.807189 4921 generic.go:334] "Generic (PLEG): container finished" podID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerID="99092045bfc50f1d94f9148a0bc80b72dab0e99bdb189811a5a3033a5fd40145" exitCode=0 Jan 03 03:53:21 crc kubenswrapper[4921]: I0103 03:53:21.807326 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerDied","Data":"99092045bfc50f1d94f9148a0bc80b72dab0e99bdb189811a5a3033a5fd40145"} Jan 03 03:53:22 crc kubenswrapper[4921]: I0103 03:53:22.816603 4921 generic.go:334] "Generic (PLEG): container finished" podID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerID="017201b0c5eb15ba1f06930657fd0ae3740122aeb223467ccb8f83c310722d31" exitCode=0 Jan 03 03:53:22 crc kubenswrapper[4921]: I0103 03:53:22.816703 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerDied","Data":"017201b0c5eb15ba1f06930657fd0ae3740122aeb223467ccb8f83c310722d31"} Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.117253 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.239803 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqpdg\" (UniqueName: \"kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg\") pod \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.240007 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util\") pod \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.240159 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle\") pod \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\" (UID: \"da345cb0-e3ff-4397-bc44-1e0e4b419c59\") " Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.241382 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle" (OuterVolumeSpecName: "bundle") pod "da345cb0-e3ff-4397-bc44-1e0e4b419c59" (UID: "da345cb0-e3ff-4397-bc44-1e0e4b419c59"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.245800 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg" (OuterVolumeSpecName: "kube-api-access-gqpdg") pod "da345cb0-e3ff-4397-bc44-1e0e4b419c59" (UID: "da345cb0-e3ff-4397-bc44-1e0e4b419c59"). InnerVolumeSpecName "kube-api-access-gqpdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.255104 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util" (OuterVolumeSpecName: "util") pod "da345cb0-e3ff-4397-bc44-1e0e4b419c59" (UID: "da345cb0-e3ff-4397-bc44-1e0e4b419c59"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.342238 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.342677 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/da345cb0-e3ff-4397-bc44-1e0e4b419c59-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.342825 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqpdg\" (UniqueName: \"kubernetes.io/projected/da345cb0-e3ff-4397-bc44-1e0e4b419c59-kube-api-access-gqpdg\") on node \"crc\" DevicePath \"\"" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.833873 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" event={"ID":"da345cb0-e3ff-4397-bc44-1e0e4b419c59","Type":"ContainerDied","Data":"9e2a8234dc652329962cbc29374dadbcb531feb466b34638067b149ce39f2a0e"} Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.834151 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e2a8234dc652329962cbc29374dadbcb531feb466b34638067b149ce39f2a0e" Jan 03 03:53:24 crc kubenswrapper[4921]: I0103 03:53:24.834220 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4" Jan 03 03:53:26 crc kubenswrapper[4921]: I0103 03:53:26.267930 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w52bw" Jan 03 03:53:31 crc kubenswrapper[4921]: I0103 03:53:31.176057 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:53:31 crc kubenswrapper[4921]: I0103 03:53:31.176322 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.435211 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v"] Jan 03 03:53:33 crc kubenswrapper[4921]: E0103 03:53:33.435531 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="util" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.435548 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="util" Jan 03 03:53:33 crc kubenswrapper[4921]: E0103 03:53:33.435560 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="extract" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.435571 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="extract" Jan 03 03:53:33 crc kubenswrapper[4921]: E0103 03:53:33.435591 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="pull" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.435598 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="pull" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.435699 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="da345cb0-e3ff-4397-bc44-1e0e4b419c59" containerName="extract" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.436347 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.438794 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.439048 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-5fgpq" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.439061 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.439176 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.439205 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.456573 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v"] Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.608170 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-webhook-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.608231 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-apiservice-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.608319 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txqs7\" (UniqueName: \"kubernetes.io/projected/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-kube-api-access-txqs7\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.710333 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txqs7\" (UniqueName: \"kubernetes.io/projected/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-kube-api-access-txqs7\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.710934 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-webhook-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.710967 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-apiservice-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.719777 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-apiservice-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.721364 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-webhook-cert\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.737150 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878"] Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.737970 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.740609 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.741215 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-b9qfw" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.742989 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.756343 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878"] Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.765416 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txqs7\" (UniqueName: \"kubernetes.io/projected/6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0-kube-api-access-txqs7\") pod \"metallb-operator-controller-manager-87945dc6b-c7r8v\" (UID: \"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0\") " pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.913920 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m6bp\" (UniqueName: \"kubernetes.io/projected/71acd931-ba27-4db9-85ef-cbe3a14d28b5-kube-api-access-2m6bp\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.913988 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-webhook-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:33 crc kubenswrapper[4921]: I0103 03:53:33.914025 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-apiservice-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.015313 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m6bp\" (UniqueName: \"kubernetes.io/projected/71acd931-ba27-4db9-85ef-cbe3a14d28b5-kube-api-access-2m6bp\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.015391 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-webhook-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.015454 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-apiservice-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.019081 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-apiservice-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.019672 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71acd931-ba27-4db9-85ef-cbe3a14d28b5-webhook-cert\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.030166 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m6bp\" (UniqueName: \"kubernetes.io/projected/71acd931-ba27-4db9-85ef-cbe3a14d28b5-kube-api-access-2m6bp\") pod \"metallb-operator-webhook-server-5587bf6f4f-lb878\" (UID: \"71acd931-ba27-4db9-85ef-cbe3a14d28b5\") " pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.051569 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.095968 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.378435 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v"] Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.425449 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878"] Jan 03 03:53:34 crc kubenswrapper[4921]: W0103 03:53:34.439112 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71acd931_ba27_4db9_85ef_cbe3a14d28b5.slice/crio-3360833470b07792d352a5bc64eb1b4d9a633871e8c9b13adb186850bfac06ff WatchSource:0}: Error finding container 3360833470b07792d352a5bc64eb1b4d9a633871e8c9b13adb186850bfac06ff: Status 404 returned error can't find the container with id 3360833470b07792d352a5bc64eb1b4d9a633871e8c9b13adb186850bfac06ff Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.904079 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" event={"ID":"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0","Type":"ContainerStarted","Data":"fecffbfb8c67980c063d451cf494a1281abbe43ed9d189127313d41641657e34"} Jan 03 03:53:34 crc kubenswrapper[4921]: I0103 03:53:34.905491 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" event={"ID":"71acd931-ba27-4db9-85ef-cbe3a14d28b5","Type":"ContainerStarted","Data":"3360833470b07792d352a5bc64eb1b4d9a633871e8c9b13adb186850bfac06ff"} Jan 03 03:53:38 crc kubenswrapper[4921]: I0103 03:53:38.033066 4921 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.946817 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" event={"ID":"71acd931-ba27-4db9-85ef-cbe3a14d28b5","Type":"ContainerStarted","Data":"42d8649f089a77f13803dd62b6b1e74b3537970b14ea5608aa16ea13bde4027b"} Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.948604 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.958934 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" event={"ID":"6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0","Type":"ContainerStarted","Data":"d7b06047e1e9903a3642ae6a60f5016a26895100e9f262817ad01340fb6f2524"} Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.960089 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.971349 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" podStartSLOduration=1.857904862 podStartE2EDuration="6.971327268s" podCreationTimestamp="2026-01-03 03:53:33 +0000 UTC" firstStartedPulling="2026-01-03 03:53:34.441790797 +0000 UTC m=+750.053217621" lastFinishedPulling="2026-01-03 03:53:39.555213203 +0000 UTC m=+755.166640027" observedRunningTime="2026-01-03 03:53:39.966973437 +0000 UTC m=+755.578400261" watchObservedRunningTime="2026-01-03 03:53:39.971327268 +0000 UTC m=+755.582754102" Jan 03 03:53:39 crc kubenswrapper[4921]: I0103 03:53:39.995958 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" podStartSLOduration=1.865333777 podStartE2EDuration="6.995928116s" podCreationTimestamp="2026-01-03 03:53:33 +0000 UTC" firstStartedPulling="2026-01-03 03:53:34.402758841 +0000 UTC m=+750.014185665" lastFinishedPulling="2026-01-03 03:53:39.53335318 +0000 UTC m=+755.144780004" observedRunningTime="2026-01-03 03:53:39.995829153 +0000 UTC m=+755.607255997" watchObservedRunningTime="2026-01-03 03:53:39.995928116 +0000 UTC m=+755.607354940" Jan 03 03:53:54 crc kubenswrapper[4921]: I0103 03:53:54.175538 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5587bf6f4f-lb878" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.720625 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.722517 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.735471 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.868130 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.868208 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.868238 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2wf6\" (UniqueName: \"kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.969112 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.969181 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.969217 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2wf6\" (UniqueName: \"kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.970806 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:56 crc kubenswrapper[4921]: I0103 03:53:56.970890 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:57 crc kubenswrapper[4921]: I0103 03:53:57.004160 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2wf6\" (UniqueName: \"kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6\") pod \"community-operators-pj5s9\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:57 crc kubenswrapper[4921]: I0103 03:53:57.043819 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:53:57 crc kubenswrapper[4921]: I0103 03:53:57.477229 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:53:58 crc kubenswrapper[4921]: I0103 03:53:58.092586 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerStarted","Data":"81c1863a1c43b46e214687cfa884cd90e44069bdd948571869577525add29ff9"} Jan 03 03:53:59 crc kubenswrapper[4921]: I0103 03:53:59.101860 4921 generic.go:334] "Generic (PLEG): container finished" podID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerID="9f96d61585bf188606146b7c18fcbffd34b499354aaf513dad3539ab046ec3ea" exitCode=0 Jan 03 03:53:59 crc kubenswrapper[4921]: I0103 03:53:59.101977 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerDied","Data":"9f96d61585bf188606146b7c18fcbffd34b499354aaf513dad3539ab046ec3ea"} Jan 03 03:54:01 crc kubenswrapper[4921]: I0103 03:54:01.175750 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:54:01 crc kubenswrapper[4921]: I0103 03:54:01.176445 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:54:01 crc kubenswrapper[4921]: I0103 03:54:01.176572 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:54:01 crc kubenswrapper[4921]: I0103 03:54:01.177693 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 03:54:01 crc kubenswrapper[4921]: I0103 03:54:01.177857 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2" gracePeriod=600 Jan 03 03:54:02 crc kubenswrapper[4921]: I0103 03:54:02.126661 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2" exitCode=0 Jan 03 03:54:02 crc kubenswrapper[4921]: I0103 03:54:02.126773 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2"} Jan 03 03:54:02 crc kubenswrapper[4921]: I0103 03:54:02.126892 4921 scope.go:117] "RemoveContainer" containerID="5d2b19ce02f4a0c728f1fbba2ed5101961e4c249b30de16d10507507992f6893" Jan 03 03:54:02 crc kubenswrapper[4921]: I0103 03:54:02.129570 4921 generic.go:334] "Generic (PLEG): container finished" podID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerID="6269f7f44e151c06cb9b24c8571c650f2608d4c883da4408d136a1577f664c69" exitCode=0 Jan 03 03:54:02 crc kubenswrapper[4921]: I0103 03:54:02.129615 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerDied","Data":"6269f7f44e151c06cb9b24c8571c650f2608d4c883da4408d136a1577f664c69"} Jan 03 03:54:03 crc kubenswrapper[4921]: I0103 03:54:03.157012 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd"} Jan 03 03:54:03 crc kubenswrapper[4921]: I0103 03:54:03.175846 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerStarted","Data":"feed782d960c25ea1ce8cee84d0db48c3ec5d789b135a729699acaa4477b5ab2"} Jan 03 03:54:03 crc kubenswrapper[4921]: I0103 03:54:03.218066 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pj5s9" podStartSLOduration=3.4328689629999998 podStartE2EDuration="7.218042123s" podCreationTimestamp="2026-01-03 03:53:56 +0000 UTC" firstStartedPulling="2026-01-03 03:53:59.104616551 +0000 UTC m=+774.716043375" lastFinishedPulling="2026-01-03 03:54:02.889789711 +0000 UTC m=+778.501216535" observedRunningTime="2026-01-03 03:54:03.216421878 +0000 UTC m=+778.827848702" watchObservedRunningTime="2026-01-03 03:54:03.218042123 +0000 UTC m=+778.829468937" Jan 03 03:54:07 crc kubenswrapper[4921]: I0103 03:54:07.044985 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:07 crc kubenswrapper[4921]: I0103 03:54:07.046057 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:07 crc kubenswrapper[4921]: I0103 03:54:07.103670 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.054364 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-87945dc6b-c7r8v" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.129307 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.130709 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.145678 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.199132 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.199191 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.199239 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47vjz\" (UniqueName: \"kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.300884 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.301186 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.301380 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47vjz\" (UniqueName: \"kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.301512 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.301543 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.325568 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47vjz\" (UniqueName: \"kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz\") pod \"redhat-marketplace-j56wg\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.456621 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.693010 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.833156 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.834236 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.836914 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-lg6jr" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.837129 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.847174 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zmp6h"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.849702 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.854913 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.855073 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.857539 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914147 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914198 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914251 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-sockets\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914289 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9dr7\" (UniqueName: \"kubernetes.io/projected/b520d64a-8e06-4c1c-abf1-e55d308aaedd-kube-api-access-v9dr7\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914309 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b520d64a-8e06-4c1c-abf1-e55d308aaedd-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914342 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-conf\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914356 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdvd\" (UniqueName: \"kubernetes.io/projected/7614f71a-a040-40d4-9f0a-9ebee98a7aed-kube-api-access-5kdvd\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914382 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-reloader\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.914420 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-startup\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.918210 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.919123 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.923871 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.968780 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hbzjk"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.969876 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hbzjk" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.974028 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.974071 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.974167 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-4hrxc" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.974205 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.986812 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-rkjb5"] Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.987579 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:14 crc kubenswrapper[4921]: I0103 03:54:14.993569 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015693 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3b5f636d-3481-4563-a5a7-a5564c024ec5-metallb-excludel2\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015759 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-sockets\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015788 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dclvn\" (UniqueName: \"kubernetes.io/projected/3b5f636d-3481-4563-a5a7-a5564c024ec5-kube-api-access-dclvn\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015828 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015855 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9dr7\" (UniqueName: \"kubernetes.io/projected/b520d64a-8e06-4c1c-abf1-e55d308aaedd-kube-api-access-v9dr7\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015880 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b520d64a-8e06-4c1c-abf1-e55d308aaedd-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015907 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015936 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rts9q\" (UniqueName: \"kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015959 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-conf\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.015978 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdvd\" (UniqueName: \"kubernetes.io/projected/7614f71a-a040-40d4-9f0a-9ebee98a7aed-kube-api-access-5kdvd\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016000 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016032 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-reloader\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016056 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016084 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-startup\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016111 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r64mv\" (UniqueName: \"kubernetes.io/projected/b9839659-8190-4224-ba78-2ef853498135-kube-api-access-r64mv\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016130 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-cert\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016153 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016177 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016199 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016671 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-sockets\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.016921 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-reloader\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.017384 4921 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.017453 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs podName:7614f71a-a040-40d4-9f0a-9ebee98a7aed nodeName:}" failed. No retries permitted until 2026-01-03 03:54:15.517435545 +0000 UTC m=+791.128862359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs") pod "frr-k8s-zmp6h" (UID: "7614f71a-a040-40d4-9f0a-9ebee98a7aed") : secret "frr-k8s-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.017624 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-conf\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.017718 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-rkjb5"] Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.018036 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/7614f71a-a040-40d4-9f0a-9ebee98a7aed-frr-startup\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.018056 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.022348 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b520d64a-8e06-4c1c-abf1-e55d308aaedd-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.033033 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdvd\" (UniqueName: \"kubernetes.io/projected/7614f71a-a040-40d4-9f0a-9ebee98a7aed-kube-api-access-5kdvd\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.035202 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9dr7\" (UniqueName: \"kubernetes.io/projected/b520d64a-8e06-4c1c-abf1-e55d308aaedd-kube-api-access-v9dr7\") pod \"frr-k8s-webhook-server-7784b6fcf-5vrrb\" (UID: \"b520d64a-8e06-4c1c-abf1-e55d308aaedd\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117342 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rts9q\" (UniqueName: \"kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117396 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117439 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117490 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r64mv\" (UniqueName: \"kubernetes.io/projected/b9839659-8190-4224-ba78-2ef853498135-kube-api-access-r64mv\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117514 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-cert\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117566 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117581 4921 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117644 4921 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117593 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3b5f636d-3481-4563-a5a7-a5564c024ec5-metallb-excludel2\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117699 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs podName:b9839659-8190-4224-ba78-2ef853498135 nodeName:}" failed. No retries permitted until 2026-01-03 03:54:15.617661633 +0000 UTC m=+791.229088457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs") pod "controller-5bddd4b946-rkjb5" (UID: "b9839659-8190-4224-ba78-2ef853498135") : secret "controller-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117722 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs podName:3b5f636d-3481-4563-a5a7-a5564c024ec5 nodeName:}" failed. No retries permitted until 2026-01-03 03:54:15.617711144 +0000 UTC m=+791.229138048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs") pod "speaker-hbzjk" (UID: "3b5f636d-3481-4563-a5a7-a5564c024ec5") : secret "speaker-certs-secret" not found Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117724 4921 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117753 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dclvn\" (UniqueName: \"kubernetes.io/projected/3b5f636d-3481-4563-a5a7-a5564c024ec5-kube-api-access-dclvn\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.117805 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist podName:3b5f636d-3481-4563-a5a7-a5564c024ec5 nodeName:}" failed. No retries permitted until 2026-01-03 03:54:15.617784836 +0000 UTC m=+791.229211660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist") pod "speaker-hbzjk" (UID: "3b5f636d-3481-4563-a5a7-a5564c024ec5") : secret "metallb-memberlist" not found Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117830 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.117877 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.118348 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3b5f636d-3481-4563-a5a7-a5564c024ec5-metallb-excludel2\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.118420 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.118444 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.119126 4921 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.131318 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-cert\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.137752 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dclvn\" (UniqueName: \"kubernetes.io/projected/3b5f636d-3481-4563-a5a7-a5564c024ec5-kube-api-access-dclvn\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.143390 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rts9q\" (UniqueName: \"kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q\") pod \"certified-operators-dws9p\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.145200 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r64mv\" (UniqueName: \"kubernetes.io/projected/b9839659-8190-4224-ba78-2ef853498135-kube-api-access-r64mv\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.156075 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.254683 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.259811 4921 generic.go:334] "Generic (PLEG): container finished" podID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerID="b96dfe05fe40242a1893c223480cbab530befc17f7239ee1f4297d6816472618" exitCode=0 Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.259846 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerDied","Data":"b96dfe05fe40242a1893c223480cbab530befc17f7239ee1f4297d6816472618"} Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.259871 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerStarted","Data":"af380c80dfaa48a06759094769fb072529633a438297f4881d529c13ee1e9ac9"} Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.522295 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.524648 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.530386 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7614f71a-a040-40d4-9f0a-9ebee98a7aed-metrics-certs\") pod \"frr-k8s-zmp6h\" (UID: \"7614f71a-a040-40d4-9f0a-9ebee98a7aed\") " pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: W0103 03:54:15.581739 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb520d64a_8e06_4c1c_abf1_e55d308aaedd.slice/crio-18c47ba9f917871cb2eab1b8a53614ce36c427cc4928d056216b76b54f0c7cf3 WatchSource:0}: Error finding container 18c47ba9f917871cb2eab1b8a53614ce36c427cc4928d056216b76b54f0c7cf3: Status 404 returned error can't find the container with id 18c47ba9f917871cb2eab1b8a53614ce36c427cc4928d056216b76b54f0c7cf3 Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.582371 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb"] Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.629525 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.629623 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.629658 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.630001 4921 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.630134 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist podName:3b5f636d-3481-4563-a5a7-a5564c024ec5 nodeName:}" failed. No retries permitted until 2026-01-03 03:54:16.630100608 +0000 UTC m=+792.241527432 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist") pod "speaker-hbzjk" (UID: "3b5f636d-3481-4563-a5a7-a5564c024ec5") : secret "metallb-memberlist" not found Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.651719 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-metrics-certs\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.655976 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b9839659-8190-4224-ba78-2ef853498135-metrics-certs\") pod \"controller-5bddd4b946-rkjb5\" (UID: \"b9839659-8190-4224-ba78-2ef853498135\") " pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.767485 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:15 crc kubenswrapper[4921]: E0103 03:54:15.805052 4921 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod554f4a02_97b9_466d_bd44_91790c56cbce.slice/crio-conmon-08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e.scope\": RecentStats: unable to find data in memory cache]" Jan 03 03:54:15 crc kubenswrapper[4921]: I0103 03:54:15.908745 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.124783 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-rkjb5"] Jan 03 03:54:16 crc kubenswrapper[4921]: W0103 03:54:16.128217 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9839659_8190_4224_ba78_2ef853498135.slice/crio-e7a537d2b1e65c92110c154b1be6637e25150c5ea54742b8fcdea51bb556560d WatchSource:0}: Error finding container e7a537d2b1e65c92110c154b1be6637e25150c5ea54742b8fcdea51bb556560d: Status 404 returned error can't find the container with id e7a537d2b1e65c92110c154b1be6637e25150c5ea54742b8fcdea51bb556560d Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.268830 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-rkjb5" event={"ID":"b9839659-8190-4224-ba78-2ef853498135","Type":"ContainerStarted","Data":"eaf6b65f0176bdc9d8b9597897ed3323c7058ca3d790ae3fc57a55b87e856bf1"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.268879 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-rkjb5" event={"ID":"b9839659-8190-4224-ba78-2ef853498135","Type":"ContainerStarted","Data":"e7a537d2b1e65c92110c154b1be6637e25150c5ea54742b8fcdea51bb556560d"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.269941 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"99da08c79de7ddd1376775f67168239f3652ec5463e6e84126e7100d6968e6ed"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.271518 4921 generic.go:334] "Generic (PLEG): container finished" podID="554f4a02-97b9-466d-bd44-91790c56cbce" containerID="08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e" exitCode=0 Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.271566 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerDied","Data":"08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.271585 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerStarted","Data":"e6f1fcc2a0aa984ab9da3506dc0a8480e38440b95c10aec42931984420d39b14"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.273611 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" event={"ID":"b520d64a-8e06-4c1c-abf1-e55d308aaedd","Type":"ContainerStarted","Data":"18c47ba9f917871cb2eab1b8a53614ce36c427cc4928d056216b76b54f0c7cf3"} Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.645992 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.652593 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3b5f636d-3481-4563-a5a7-a5564c024ec5-memberlist\") pod \"speaker-hbzjk\" (UID: \"3b5f636d-3481-4563-a5a7-a5564c024ec5\") " pod="metallb-system/speaker-hbzjk" Jan 03 03:54:16 crc kubenswrapper[4921]: I0103 03:54:16.790022 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hbzjk" Jan 03 03:54:16 crc kubenswrapper[4921]: W0103 03:54:16.833498 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b5f636d_3481_4563_a5a7_a5564c024ec5.slice/crio-9d3f8839459a980944d80da5effeba9271ca4719bf9ac7c0ff7fa18a5953ed95 WatchSource:0}: Error finding container 9d3f8839459a980944d80da5effeba9271ca4719bf9ac7c0ff7fa18a5953ed95: Status 404 returned error can't find the container with id 9d3f8839459a980944d80da5effeba9271ca4719bf9ac7c0ff7fa18a5953ed95 Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.132218 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.291026 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hbzjk" event={"ID":"3b5f636d-3481-4563-a5a7-a5564c024ec5","Type":"ContainerStarted","Data":"22078aaa90c83a0b7be01d4c283e87f2a597812f3c013a6a99688795c7408ab6"} Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.291063 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hbzjk" event={"ID":"3b5f636d-3481-4563-a5a7-a5564c024ec5","Type":"ContainerStarted","Data":"9d3f8839459a980944d80da5effeba9271ca4719bf9ac7c0ff7fa18a5953ed95"} Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.296935 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerStarted","Data":"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4"} Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.298898 4921 generic.go:334] "Generic (PLEG): container finished" podID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerID="558b8ca77f0bf1ec5194cdeb3b74c70c9b0fd51126525189592fb9882c27b70a" exitCode=0 Jan 03 03:54:17 crc kubenswrapper[4921]: I0103 03:54:17.298940 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerDied","Data":"558b8ca77f0bf1ec5194cdeb3b74c70c9b0fd51126525189592fb9882c27b70a"} Jan 03 03:54:18 crc kubenswrapper[4921]: I0103 03:54:18.314646 4921 generic.go:334] "Generic (PLEG): container finished" podID="554f4a02-97b9-466d-bd44-91790c56cbce" containerID="7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4" exitCode=0 Jan 03 03:54:18 crc kubenswrapper[4921]: I0103 03:54:18.314785 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerDied","Data":"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4"} Jan 03 03:54:18 crc kubenswrapper[4921]: I0103 03:54:18.319513 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerStarted","Data":"0f3ce12676b8f897e731ba67ade806376041b68570b605decbe6ce680c97b65f"} Jan 03 03:54:18 crc kubenswrapper[4921]: I0103 03:54:18.355255 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j56wg" podStartSLOduration=1.809526888 podStartE2EDuration="4.355239077s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:15.267008953 +0000 UTC m=+790.878435777" lastFinishedPulling="2026-01-03 03:54:17.812721142 +0000 UTC m=+793.424147966" observedRunningTime="2026-01-03 03:54:18.355031792 +0000 UTC m=+793.966458636" watchObservedRunningTime="2026-01-03 03:54:18.355239077 +0000 UTC m=+793.966665901" Jan 03 03:54:19 crc kubenswrapper[4921]: I0103 03:54:19.328559 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerStarted","Data":"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1"} Jan 03 03:54:19 crc kubenswrapper[4921]: I0103 03:54:19.349851 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dws9p" podStartSLOduration=2.860851473 podStartE2EDuration="5.34983958s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:16.273330653 +0000 UTC m=+791.884757477" lastFinishedPulling="2026-01-03 03:54:18.76231876 +0000 UTC m=+794.373745584" observedRunningTime="2026-01-03 03:54:19.346634665 +0000 UTC m=+794.958061489" watchObservedRunningTime="2026-01-03 03:54:19.34983958 +0000 UTC m=+794.961266404" Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.342509 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hbzjk" event={"ID":"3b5f636d-3481-4563-a5a7-a5564c024ec5","Type":"ContainerStarted","Data":"b4916d348b07f2a864872b8d17b5c9fc65b4c9670142c773f3203bf427b1875e"} Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.342850 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hbzjk" Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.344063 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-rkjb5" event={"ID":"b9839659-8190-4224-ba78-2ef853498135","Type":"ContainerStarted","Data":"7e1c772e708c9cff16eb7c5237385faafc91d89fab0577731b3ee18eb40bc4ec"} Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.344263 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.364083 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hbzjk" podStartSLOduration=4.346444507 podStartE2EDuration="7.364060703s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:17.154524572 +0000 UTC m=+792.765951396" lastFinishedPulling="2026-01-03 03:54:20.172140768 +0000 UTC m=+795.783567592" observedRunningTime="2026-01-03 03:54:21.361723771 +0000 UTC m=+796.973150625" watchObservedRunningTime="2026-01-03 03:54:21.364060703 +0000 UTC m=+796.975487547" Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.379296 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-rkjb5" podStartSLOduration=3.475736244 podStartE2EDuration="7.379253342s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:16.251867109 +0000 UTC m=+791.863293933" lastFinishedPulling="2026-01-03 03:54:20.155384207 +0000 UTC m=+795.766811031" observedRunningTime="2026-01-03 03:54:21.376979712 +0000 UTC m=+796.988406586" watchObservedRunningTime="2026-01-03 03:54:21.379253342 +0000 UTC m=+796.990680166" Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.518338 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:54:21 crc kubenswrapper[4921]: I0103 03:54:21.518694 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pj5s9" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="registry-server" containerID="cri-o://feed782d960c25ea1ce8cee84d0db48c3ec5d789b135a729699acaa4477b5ab2" gracePeriod=2 Jan 03 03:54:22 crc kubenswrapper[4921]: I0103 03:54:22.355449 4921 generic.go:334] "Generic (PLEG): container finished" podID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerID="feed782d960c25ea1ce8cee84d0db48c3ec5d789b135a729699acaa4477b5ab2" exitCode=0 Jan 03 03:54:22 crc kubenswrapper[4921]: I0103 03:54:22.355551 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerDied","Data":"feed782d960c25ea1ce8cee84d0db48c3ec5d789b135a729699acaa4477b5ab2"} Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.457710 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.458007 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.507829 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.565411 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.731210 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2wf6\" (UniqueName: \"kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6\") pod \"e55dfac9-6bc1-4908-9235-080f953a7b14\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.731302 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities\") pod \"e55dfac9-6bc1-4908-9235-080f953a7b14\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.731428 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content\") pod \"e55dfac9-6bc1-4908-9235-080f953a7b14\" (UID: \"e55dfac9-6bc1-4908-9235-080f953a7b14\") " Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.732211 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities" (OuterVolumeSpecName: "utilities") pod "e55dfac9-6bc1-4908-9235-080f953a7b14" (UID: "e55dfac9-6bc1-4908-9235-080f953a7b14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.745248 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6" (OuterVolumeSpecName: "kube-api-access-d2wf6") pod "e55dfac9-6bc1-4908-9235-080f953a7b14" (UID: "e55dfac9-6bc1-4908-9235-080f953a7b14"). InnerVolumeSpecName "kube-api-access-d2wf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.780270 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e55dfac9-6bc1-4908-9235-080f953a7b14" (UID: "e55dfac9-6bc1-4908-9235-080f953a7b14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.832755 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2wf6\" (UniqueName: \"kubernetes.io/projected/e55dfac9-6bc1-4908-9235-080f953a7b14-kube-api-access-d2wf6\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.832804 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:24 crc kubenswrapper[4921]: I0103 03:54:24.832821 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e55dfac9-6bc1-4908-9235-080f953a7b14-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.255789 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.255867 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.308595 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.379227 4921 generic.go:334] "Generic (PLEG): container finished" podID="7614f71a-a040-40d4-9f0a-9ebee98a7aed" containerID="8512efadc16c75819eb631721e59cf812ad660042b0b44c3ac1a4feead37488a" exitCode=0 Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.379321 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerDied","Data":"8512efadc16c75819eb631721e59cf812ad660042b0b44c3ac1a4feead37488a"} Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.381733 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pj5s9" event={"ID":"e55dfac9-6bc1-4908-9235-080f953a7b14","Type":"ContainerDied","Data":"81c1863a1c43b46e214687cfa884cd90e44069bdd948571869577525add29ff9"} Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.381765 4921 scope.go:117] "RemoveContainer" containerID="feed782d960c25ea1ce8cee84d0db48c3ec5d789b135a729699acaa4477b5ab2" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.381861 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pj5s9" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.388486 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" event={"ID":"b520d64a-8e06-4c1c-abf1-e55d308aaedd","Type":"ContainerStarted","Data":"835c4c00435893fb698ef334a82570bd4ed2346560c2a222ad7decdf2a91ed72"} Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.418046 4921 scope.go:117] "RemoveContainer" containerID="6269f7f44e151c06cb9b24c8571c650f2608d4c883da4408d136a1577f664c69" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.433212 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.438594 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pj5s9"] Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.441371 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.449344 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.454581 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" podStartSLOduration=2.617929851 podStartE2EDuration="11.454539131s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:15.584557688 +0000 UTC m=+791.195984512" lastFinishedPulling="2026-01-03 03:54:24.421166968 +0000 UTC m=+800.032593792" observedRunningTime="2026-01-03 03:54:25.452397954 +0000 UTC m=+801.063824798" watchObservedRunningTime="2026-01-03 03:54:25.454539131 +0000 UTC m=+801.065965965" Jan 03 03:54:25 crc kubenswrapper[4921]: I0103 03:54:25.467735 4921 scope.go:117] "RemoveContainer" containerID="9f96d61585bf188606146b7c18fcbffd34b499354aaf513dad3539ab046ec3ea" Jan 03 03:54:26 crc kubenswrapper[4921]: I0103 03:54:26.396726 4921 generic.go:334] "Generic (PLEG): container finished" podID="7614f71a-a040-40d4-9f0a-9ebee98a7aed" containerID="1bb74ab59906375c757c6e12f0125c605a4d2b55cfccb54b793760e81cc19e44" exitCode=0 Jan 03 03:54:26 crc kubenswrapper[4921]: I0103 03:54:26.396797 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerDied","Data":"1bb74ab59906375c757c6e12f0125c605a4d2b55cfccb54b793760e81cc19e44"} Jan 03 03:54:26 crc kubenswrapper[4921]: I0103 03:54:26.399131 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:26 crc kubenswrapper[4921]: I0103 03:54:26.892147 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" path="/var/lib/kubelet/pods/e55dfac9-6bc1-4908-9235-080f953a7b14/volumes" Jan 03 03:54:27 crc kubenswrapper[4921]: I0103 03:54:27.407839 4921 generic.go:334] "Generic (PLEG): container finished" podID="7614f71a-a040-40d4-9f0a-9ebee98a7aed" containerID="166e227838431691b6dc3ea5344f0e904fd08ada7e54f14de554824132ca0333" exitCode=0 Jan 03 03:54:27 crc kubenswrapper[4921]: I0103 03:54:27.407909 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerDied","Data":"166e227838431691b6dc3ea5344f0e904fd08ada7e54f14de554824132ca0333"} Jan 03 03:54:28 crc kubenswrapper[4921]: I0103 03:54:28.110624 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:28 crc kubenswrapper[4921]: I0103 03:54:28.111563 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j56wg" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="registry-server" containerID="cri-o://0f3ce12676b8f897e731ba67ade806376041b68570b605decbe6ce680c97b65f" gracePeriod=2 Jan 03 03:54:28 crc kubenswrapper[4921]: I0103 03:54:28.430695 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"b18a6bedb76e862edfcb021226d8fb77a4da4d2c769df17c731b0e086a3b5a7e"} Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.441174 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"514530455a0724eded3cd83ae2e2ed4237443ee3fab91120558bd1d03c862643"} Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.443351 4921 generic.go:334] "Generic (PLEG): container finished" podID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerID="0f3ce12676b8f897e731ba67ade806376041b68570b605decbe6ce680c97b65f" exitCode=0 Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.443394 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerDied","Data":"0f3ce12676b8f897e731ba67ade806376041b68570b605decbe6ce680c97b65f"} Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.705199 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.709264 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.709623 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dws9p" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="registry-server" containerID="cri-o://cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1" gracePeriod=2 Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.806133 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content\") pod \"91e35f39-1a47-463b-9752-55f58aa9bd21\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.806208 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47vjz\" (UniqueName: \"kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz\") pod \"91e35f39-1a47-463b-9752-55f58aa9bd21\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.806238 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities\") pod \"91e35f39-1a47-463b-9752-55f58aa9bd21\" (UID: \"91e35f39-1a47-463b-9752-55f58aa9bd21\") " Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.807348 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities" (OuterVolumeSpecName: "utilities") pod "91e35f39-1a47-463b-9752-55f58aa9bd21" (UID: "91e35f39-1a47-463b-9752-55f58aa9bd21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.811326 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz" (OuterVolumeSpecName: "kube-api-access-47vjz") pod "91e35f39-1a47-463b-9752-55f58aa9bd21" (UID: "91e35f39-1a47-463b-9752-55f58aa9bd21"). InnerVolumeSpecName "kube-api-access-47vjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.828256 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91e35f39-1a47-463b-9752-55f58aa9bd21" (UID: "91e35f39-1a47-463b-9752-55f58aa9bd21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.907653 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.907695 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47vjz\" (UniqueName: \"kubernetes.io/projected/91e35f39-1a47-463b-9752-55f58aa9bd21-kube-api-access-47vjz\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:29 crc kubenswrapper[4921]: I0103 03:54:29.907709 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35f39-1a47-463b-9752-55f58aa9bd21-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.019414 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.109678 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content\") pod \"554f4a02-97b9-466d-bd44-91790c56cbce\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.109747 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities\") pod \"554f4a02-97b9-466d-bd44-91790c56cbce\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.109822 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rts9q\" (UniqueName: \"kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q\") pod \"554f4a02-97b9-466d-bd44-91790c56cbce\" (UID: \"554f4a02-97b9-466d-bd44-91790c56cbce\") " Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.110909 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities" (OuterVolumeSpecName: "utilities") pod "554f4a02-97b9-466d-bd44-91790c56cbce" (UID: "554f4a02-97b9-466d-bd44-91790c56cbce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.112935 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q" (OuterVolumeSpecName: "kube-api-access-rts9q") pod "554f4a02-97b9-466d-bd44-91790c56cbce" (UID: "554f4a02-97b9-466d-bd44-91790c56cbce"). InnerVolumeSpecName "kube-api-access-rts9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.164410 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "554f4a02-97b9-466d-bd44-91790c56cbce" (UID: "554f4a02-97b9-466d-bd44-91790c56cbce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.211258 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.211307 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rts9q\" (UniqueName: \"kubernetes.io/projected/554f4a02-97b9-466d-bd44-91790c56cbce-kube-api-access-rts9q\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.211317 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554f4a02-97b9-466d-bd44-91790c56cbce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.452189 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"6d8c9e97ce3e34a5bd610735afe60da1351a6e0bf72bdc293fc44a4585212b5f"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.452496 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"567447dd140e699dd1c4f35141eec836caf6891e87eb420d05fe5f8508dbdabc"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.452508 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"6aef66f409475bbe8fd4dbd4ef3e85112eb2ee3314e3f668a829168a4147706e"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.452516 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmp6h" event={"ID":"7614f71a-a040-40d4-9f0a-9ebee98a7aed","Type":"ContainerStarted","Data":"7579c1ba899366e0a0bd2e76b4f7689546bb290601fbd3a37db8fca1ef247b9a"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.453453 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.455627 4921 generic.go:334] "Generic (PLEG): container finished" podID="554f4a02-97b9-466d-bd44-91790c56cbce" containerID="cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1" exitCode=0 Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.455677 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerDied","Data":"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.455767 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dws9p" event={"ID":"554f4a02-97b9-466d-bd44-91790c56cbce","Type":"ContainerDied","Data":"e6f1fcc2a0aa984ab9da3506dc0a8480e38440b95c10aec42931984420d39b14"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.455794 4921 scope.go:117] "RemoveContainer" containerID="cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.455772 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dws9p" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.460932 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j56wg" event={"ID":"91e35f39-1a47-463b-9752-55f58aa9bd21","Type":"ContainerDied","Data":"af380c80dfaa48a06759094769fb072529633a438297f4881d529c13ee1e9ac9"} Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.461051 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j56wg" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.477670 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zmp6h" podStartSLOduration=7.944379813 podStartE2EDuration="16.47765232s" podCreationTimestamp="2026-01-03 03:54:14 +0000 UTC" firstStartedPulling="2026-01-03 03:54:15.879060989 +0000 UTC m=+791.490487813" lastFinishedPulling="2026-01-03 03:54:24.412333486 +0000 UTC m=+800.023760320" observedRunningTime="2026-01-03 03:54:30.47652452 +0000 UTC m=+806.087951394" watchObservedRunningTime="2026-01-03 03:54:30.47765232 +0000 UTC m=+806.089079144" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.482745 4921 scope.go:117] "RemoveContainer" containerID="7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.511386 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.517331 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j56wg"] Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.548442 4921 scope.go:117] "RemoveContainer" containerID="08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.548472 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.554715 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dws9p"] Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.589489 4921 scope.go:117] "RemoveContainer" containerID="cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1" Jan 03 03:54:30 crc kubenswrapper[4921]: E0103 03:54:30.593409 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1\": container with ID starting with cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1 not found: ID does not exist" containerID="cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.593465 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1"} err="failed to get container status \"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1\": rpc error: code = NotFound desc = could not find container \"cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1\": container with ID starting with cf0f07075a3650d547dbf51b03ce3bd8d1328a8c3d02af503a09879f6201f3b1 not found: ID does not exist" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.593565 4921 scope.go:117] "RemoveContainer" containerID="7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4" Jan 03 03:54:30 crc kubenswrapper[4921]: E0103 03:54:30.599524 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4\": container with ID starting with 7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4 not found: ID does not exist" containerID="7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.599584 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4"} err="failed to get container status \"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4\": rpc error: code = NotFound desc = could not find container \"7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4\": container with ID starting with 7dd4dfadb675d8a56297458ae8f08f6be5d8761c9af30b445f5a70440ba2a5d4 not found: ID does not exist" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.599616 4921 scope.go:117] "RemoveContainer" containerID="08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e" Jan 03 03:54:30 crc kubenswrapper[4921]: E0103 03:54:30.602177 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e\": container with ID starting with 08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e not found: ID does not exist" containerID="08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.602208 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e"} err="failed to get container status \"08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e\": rpc error: code = NotFound desc = could not find container \"08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e\": container with ID starting with 08cb9559f531914ee2a449153220b78b8b4f051759b81d44d402ea588d244f3e not found: ID does not exist" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.602225 4921 scope.go:117] "RemoveContainer" containerID="0f3ce12676b8f897e731ba67ade806376041b68570b605decbe6ce680c97b65f" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.616830 4921 scope.go:117] "RemoveContainer" containerID="558b8ca77f0bf1ec5194cdeb3b74c70c9b0fd51126525189592fb9882c27b70a" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.630628 4921 scope.go:117] "RemoveContainer" containerID="b96dfe05fe40242a1893c223480cbab530befc17f7239ee1f4297d6816472618" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.768150 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.810566 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.890599 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" path="/var/lib/kubelet/pods/554f4a02-97b9-466d-bd44-91790c56cbce/volumes" Jan 03 03:54:30 crc kubenswrapper[4921]: I0103 03:54:30.891888 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" path="/var/lib/kubelet/pods/91e35f39-1a47-463b-9752-55f58aa9bd21/volumes" Jan 03 03:54:35 crc kubenswrapper[4921]: I0103 03:54:35.167791 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5vrrb" Jan 03 03:54:35 crc kubenswrapper[4921]: I0103 03:54:35.911942 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-rkjb5" Jan 03 03:54:36 crc kubenswrapper[4921]: I0103 03:54:36.794705 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hbzjk" Jan 03 03:54:45 crc kubenswrapper[4921]: I0103 03:54:45.774151 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zmp6h" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.716445 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-4cq2f"] Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.716908 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.716922 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.716939 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.716947 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.716959 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.716966 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.716974 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.716983 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.717000 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717008 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="extract-utilities" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.717020 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717029 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.717039 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717048 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="extract-content" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.717059 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717066 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: E0103 03:54:46.717081 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717089 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717212 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="554f4a02-97b9-466d-bd44-91790c56cbce" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717226 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e35f39-1a47-463b-9752-55f58aa9bd21" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717241 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="e55dfac9-6bc1-4908-9235-080f953a7b14" containerName="registry-server" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.717682 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.720025 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.720455 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-t5knb" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.720622 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.731713 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4cq2f"] Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.852874 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9cxf\" (UniqueName: \"kubernetes.io/projected/7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba-kube-api-access-m9cxf\") pod \"mariadb-operator-index-4cq2f\" (UID: \"7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba\") " pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.955850 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9cxf\" (UniqueName: \"kubernetes.io/projected/7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba-kube-api-access-m9cxf\") pod \"mariadb-operator-index-4cq2f\" (UID: \"7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba\") " pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:46 crc kubenswrapper[4921]: I0103 03:54:46.981560 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9cxf\" (UniqueName: \"kubernetes.io/projected/7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba-kube-api-access-m9cxf\") pod \"mariadb-operator-index-4cq2f\" (UID: \"7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba\") " pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:47 crc kubenswrapper[4921]: I0103 03:54:47.080018 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:47 crc kubenswrapper[4921]: I0103 03:54:47.352874 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-4cq2f"] Jan 03 03:54:47 crc kubenswrapper[4921]: I0103 03:54:47.576571 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4cq2f" event={"ID":"7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba","Type":"ContainerStarted","Data":"a925d7b60311f864394b9d04fe92cc31918b111c52c309a8576fd1546d26ad25"} Jan 03 03:54:55 crc kubenswrapper[4921]: I0103 03:54:55.632740 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-4cq2f" event={"ID":"7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba","Type":"ContainerStarted","Data":"f696eb42663f26a02166e0d10c08ed9e49d0fd145b42cc2beb1bb297c7804ec7"} Jan 03 03:54:55 crc kubenswrapper[4921]: I0103 03:54:55.651292 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-4cq2f" podStartSLOduration=1.552573674 podStartE2EDuration="9.651261816s" podCreationTimestamp="2026-01-03 03:54:46 +0000 UTC" firstStartedPulling="2026-01-03 03:54:47.359914834 +0000 UTC m=+822.971341658" lastFinishedPulling="2026-01-03 03:54:55.458602976 +0000 UTC m=+831.070029800" observedRunningTime="2026-01-03 03:54:55.649831078 +0000 UTC m=+831.261257922" watchObservedRunningTime="2026-01-03 03:54:55.651261816 +0000 UTC m=+831.262688640" Jan 03 03:54:57 crc kubenswrapper[4921]: I0103 03:54:57.081029 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:57 crc kubenswrapper[4921]: I0103 03:54:57.081363 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:54:57 crc kubenswrapper[4921]: I0103 03:54:57.117545 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:55:07 crc kubenswrapper[4921]: I0103 03:55:07.128631 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-4cq2f" Jan 03 03:55:13 crc kubenswrapper[4921]: I0103 03:55:13.982971 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p"] Jan 03 03:55:13 crc kubenswrapper[4921]: I0103 03:55:13.985987 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:13 crc kubenswrapper[4921]: I0103 03:55:13.988537 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 03:55:13 crc kubenswrapper[4921]: I0103 03:55:13.998156 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p"] Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.038643 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.038755 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwr2b\" (UniqueName: \"kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.038868 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.140520 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.140598 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwr2b\" (UniqueName: \"kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.140658 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.141206 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.141365 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.167943 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwr2b\" (UniqueName: \"kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b\") pod \"798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.322604 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:14 crc kubenswrapper[4921]: I0103 03:55:14.810718 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p"] Jan 03 03:55:15 crc kubenswrapper[4921]: I0103 03:55:15.790459 4921 generic.go:334] "Generic (PLEG): container finished" podID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerID="2899f66d163c0c6c0df0ec7797718803805beae9035bfaefa1312120b8322a9f" exitCode=0 Jan 03 03:55:15 crc kubenswrapper[4921]: I0103 03:55:15.790549 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" event={"ID":"a4c5e860-d6d5-4b05-965a-042db7d97a45","Type":"ContainerDied","Data":"2899f66d163c0c6c0df0ec7797718803805beae9035bfaefa1312120b8322a9f"} Jan 03 03:55:15 crc kubenswrapper[4921]: I0103 03:55:15.792683 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" event={"ID":"a4c5e860-d6d5-4b05-965a-042db7d97a45","Type":"ContainerStarted","Data":"dfdcefc531c8aa124b3245762d260adf5f86765e8189f00bc3e0f4b72b8d36c8"} Jan 03 03:55:16 crc kubenswrapper[4921]: I0103 03:55:16.801903 4921 generic.go:334] "Generic (PLEG): container finished" podID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerID="477e74ec852e64a6ced281c6f31368629f85002857fc930392a85038a0e9e13a" exitCode=0 Jan 03 03:55:16 crc kubenswrapper[4921]: I0103 03:55:16.802212 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" event={"ID":"a4c5e860-d6d5-4b05-965a-042db7d97a45","Type":"ContainerDied","Data":"477e74ec852e64a6ced281c6f31368629f85002857fc930392a85038a0e9e13a"} Jan 03 03:55:17 crc kubenswrapper[4921]: I0103 03:55:17.813382 4921 generic.go:334] "Generic (PLEG): container finished" podID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerID="ea865a8513172e609916deb7218892287206bf40885ac569e83ebd1441564143" exitCode=0 Jan 03 03:55:17 crc kubenswrapper[4921]: I0103 03:55:17.813476 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" event={"ID":"a4c5e860-d6d5-4b05-965a-042db7d97a45","Type":"ContainerDied","Data":"ea865a8513172e609916deb7218892287206bf40885ac569e83ebd1441564143"} Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.123864 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.216160 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwr2b\" (UniqueName: \"kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b\") pod \"a4c5e860-d6d5-4b05-965a-042db7d97a45\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.216242 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle\") pod \"a4c5e860-d6d5-4b05-965a-042db7d97a45\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.216366 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util\") pod \"a4c5e860-d6d5-4b05-965a-042db7d97a45\" (UID: \"a4c5e860-d6d5-4b05-965a-042db7d97a45\") " Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.217133 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle" (OuterVolumeSpecName: "bundle") pod "a4c5e860-d6d5-4b05-965a-042db7d97a45" (UID: "a4c5e860-d6d5-4b05-965a-042db7d97a45"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.222983 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b" (OuterVolumeSpecName: "kube-api-access-lwr2b") pod "a4c5e860-d6d5-4b05-965a-042db7d97a45" (UID: "a4c5e860-d6d5-4b05-965a-042db7d97a45"). InnerVolumeSpecName "kube-api-access-lwr2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.231054 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util" (OuterVolumeSpecName: "util") pod "a4c5e860-d6d5-4b05-965a-042db7d97a45" (UID: "a4c5e860-d6d5-4b05-965a-042db7d97a45"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.318395 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.318438 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwr2b\" (UniqueName: \"kubernetes.io/projected/a4c5e860-d6d5-4b05-965a-042db7d97a45-kube-api-access-lwr2b\") on node \"crc\" DevicePath \"\"" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.318452 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a4c5e860-d6d5-4b05-965a-042db7d97a45-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.831250 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" event={"ID":"a4c5e860-d6d5-4b05-965a-042db7d97a45","Type":"ContainerDied","Data":"dfdcefc531c8aa124b3245762d260adf5f86765e8189f00bc3e0f4b72b8d36c8"} Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.831378 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfdcefc531c8aa124b3245762d260adf5f86765e8189f00bc3e0f4b72b8d36c8" Jan 03 03:55:19 crc kubenswrapper[4921]: I0103 03:55:19.831402 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.301294 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm"] Jan 03 03:55:28 crc kubenswrapper[4921]: E0103 03:55:28.302065 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="pull" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.302082 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="pull" Jan 03 03:55:28 crc kubenswrapper[4921]: E0103 03:55:28.302098 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="util" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.302105 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="util" Jan 03 03:55:28 crc kubenswrapper[4921]: E0103 03:55:28.302124 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="extract" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.302130 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="extract" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.302227 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4c5e860-d6d5-4b05-965a-042db7d97a45" containerName="extract" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.302986 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.304623 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.304630 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.305757 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-jrg52" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.316605 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm"] Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.396103 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-webhook-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.396161 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45lzx\" (UniqueName: \"kubernetes.io/projected/63024a63-5cbc-4f16-af95-cd2c087b42dd-kube-api-access-45lzx\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.396186 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-apiservice-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.497707 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-webhook-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.497767 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45lzx\" (UniqueName: \"kubernetes.io/projected/63024a63-5cbc-4f16-af95-cd2c087b42dd-kube-api-access-45lzx\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.497797 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-apiservice-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.504239 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-apiservice-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.505113 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/63024a63-5cbc-4f16-af95-cd2c087b42dd-webhook-cert\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.512869 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45lzx\" (UniqueName: \"kubernetes.io/projected/63024a63-5cbc-4f16-af95-cd2c087b42dd-kube-api-access-45lzx\") pod \"mariadb-operator-controller-manager-677896c6f-dsgnm\" (UID: \"63024a63-5cbc-4f16-af95-cd2c087b42dd\") " pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.622807 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.867907 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm"] Jan 03 03:55:28 crc kubenswrapper[4921]: I0103 03:55:28.901105 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" event={"ID":"63024a63-5cbc-4f16-af95-cd2c087b42dd","Type":"ContainerStarted","Data":"dfadc9a80233d7468fc130c9ea45b323fe473ecf3f6db3d6f38a76bd2938107c"} Jan 03 03:55:34 crc kubenswrapper[4921]: I0103 03:55:34.962285 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" event={"ID":"63024a63-5cbc-4f16-af95-cd2c087b42dd","Type":"ContainerStarted","Data":"4a57a47632691200b73be1cedf28cd5f4e38605aa4b5833a9b695636a14cf5ed"} Jan 03 03:55:36 crc kubenswrapper[4921]: I0103 03:55:36.983393 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" event={"ID":"63024a63-5cbc-4f16-af95-cd2c087b42dd","Type":"ContainerStarted","Data":"ecfba5db82167d918f255497797dfa47c18d5f30572f62ae5d48d14121229cae"} Jan 03 03:55:36 crc kubenswrapper[4921]: I0103 03:55:36.984926 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:37 crc kubenswrapper[4921]: I0103 03:55:37.016854 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" podStartSLOduration=1.308093178 podStartE2EDuration="9.016828548s" podCreationTimestamp="2026-01-03 03:55:28 +0000 UTC" firstStartedPulling="2026-01-03 03:55:28.874913352 +0000 UTC m=+864.486340166" lastFinishedPulling="2026-01-03 03:55:36.583648702 +0000 UTC m=+872.195075536" observedRunningTime="2026-01-03 03:55:37.002080785 +0000 UTC m=+872.613507609" watchObservedRunningTime="2026-01-03 03:55:37.016828548 +0000 UTC m=+872.628255372" Jan 03 03:55:48 crc kubenswrapper[4921]: I0103 03:55:48.629093 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-677896c6f-dsgnm" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.186577 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.187409 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.189961 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-jtf6n" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.201152 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.229539 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c77z\" (UniqueName: \"kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z\") pod \"infra-operator-index-wdqwl\" (UID: \"162ffb8b-a12f-45bf-84ef-0cdcb1668673\") " pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.331020 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c77z\" (UniqueName: \"kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z\") pod \"infra-operator-index-wdqwl\" (UID: \"162ffb8b-a12f-45bf-84ef-0cdcb1668673\") " pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.357923 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c77z\" (UniqueName: \"kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z\") pod \"infra-operator-index-wdqwl\" (UID: \"162ffb8b-a12f-45bf-84ef-0cdcb1668673\") " pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:51 crc kubenswrapper[4921]: I0103 03:55:51.522627 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:52 crc kubenswrapper[4921]: I0103 03:55:52.068540 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:55:52 crc kubenswrapper[4921]: W0103 03:55:52.075553 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod162ffb8b_a12f_45bf_84ef_0cdcb1668673.slice/crio-91dcc5660665fe83061fa49fffdc3e920a4f696c24a6faf9bf13b18bdc00ded9 WatchSource:0}: Error finding container 91dcc5660665fe83061fa49fffdc3e920a4f696c24a6faf9bf13b18bdc00ded9: Status 404 returned error can't find the container with id 91dcc5660665fe83061fa49fffdc3e920a4f696c24a6faf9bf13b18bdc00ded9 Jan 03 03:55:52 crc kubenswrapper[4921]: I0103 03:55:52.100525 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wdqwl" event={"ID":"162ffb8b-a12f-45bf-84ef-0cdcb1668673","Type":"ContainerStarted","Data":"91dcc5660665fe83061fa49fffdc3e920a4f696c24a6faf9bf13b18bdc00ded9"} Jan 03 03:55:54 crc kubenswrapper[4921]: I0103 03:55:54.117377 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wdqwl" event={"ID":"162ffb8b-a12f-45bf-84ef-0cdcb1668673","Type":"ContainerStarted","Data":"ea70a24eac1f866a1b39acccf9d7a6bd5a325a1e5fe27711a49dd5c68beab50b"} Jan 03 03:55:54 crc kubenswrapper[4921]: I0103 03:55:54.135721 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-wdqwl" podStartSLOduration=2.197662903 podStartE2EDuration="3.135704435s" podCreationTimestamp="2026-01-03 03:55:51 +0000 UTC" firstStartedPulling="2026-01-03 03:55:52.079747871 +0000 UTC m=+887.691174695" lastFinishedPulling="2026-01-03 03:55:53.017789383 +0000 UTC m=+888.629216227" observedRunningTime="2026-01-03 03:55:54.130209334 +0000 UTC m=+889.741636148" watchObservedRunningTime="2026-01-03 03:55:54.135704435 +0000 UTC m=+889.747131259" Jan 03 03:55:54 crc kubenswrapper[4921]: I0103 03:55:54.976215 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.588367 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-ltmf9"] Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.589338 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.601822 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ltmf9"] Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.688256 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm95l\" (UniqueName: \"kubernetes.io/projected/59f09088-e73a-4017-ba74-9e208902a0ac-kube-api-access-tm95l\") pod \"infra-operator-index-ltmf9\" (UID: \"59f09088-e73a-4017-ba74-9e208902a0ac\") " pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.789572 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm95l\" (UniqueName: \"kubernetes.io/projected/59f09088-e73a-4017-ba74-9e208902a0ac-kube-api-access-tm95l\") pod \"infra-operator-index-ltmf9\" (UID: \"59f09088-e73a-4017-ba74-9e208902a0ac\") " pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.812501 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm95l\" (UniqueName: \"kubernetes.io/projected/59f09088-e73a-4017-ba74-9e208902a0ac-kube-api-access-tm95l\") pod \"infra-operator-index-ltmf9\" (UID: \"59f09088-e73a-4017-ba74-9e208902a0ac\") " pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:55:55 crc kubenswrapper[4921]: I0103 03:55:55.919222 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:55:56 crc kubenswrapper[4921]: I0103 03:55:56.136771 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-wdqwl" podUID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" containerName="registry-server" containerID="cri-o://ea70a24eac1f866a1b39acccf9d7a6bd5a325a1e5fe27711a49dd5c68beab50b" gracePeriod=2 Jan 03 03:55:56 crc kubenswrapper[4921]: I0103 03:55:56.260187 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ltmf9"] Jan 03 03:55:56 crc kubenswrapper[4921]: W0103 03:55:56.294765 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59f09088_e73a_4017_ba74_9e208902a0ac.slice/crio-996eba322e169b80645f26c8f858859571c8652e05e870025b5304398ee938b4 WatchSource:0}: Error finding container 996eba322e169b80645f26c8f858859571c8652e05e870025b5304398ee938b4: Status 404 returned error can't find the container with id 996eba322e169b80645f26c8f858859571c8652e05e870025b5304398ee938b4 Jan 03 03:55:57 crc kubenswrapper[4921]: I0103 03:55:57.142430 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ltmf9" event={"ID":"59f09088-e73a-4017-ba74-9e208902a0ac","Type":"ContainerStarted","Data":"996eba322e169b80645f26c8f858859571c8652e05e870025b5304398ee938b4"} Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.151120 4921 generic.go:334] "Generic (PLEG): container finished" podID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" containerID="ea70a24eac1f866a1b39acccf9d7a6bd5a325a1e5fe27711a49dd5c68beab50b" exitCode=0 Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.151219 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wdqwl" event={"ID":"162ffb8b-a12f-45bf-84ef-0cdcb1668673","Type":"ContainerDied","Data":"ea70a24eac1f866a1b39acccf9d7a6bd5a325a1e5fe27711a49dd5c68beab50b"} Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.153167 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ltmf9" event={"ID":"59f09088-e73a-4017-ba74-9e208902a0ac","Type":"ContainerStarted","Data":"a00f95b82ea376dfe82ca80a75583c05298674f2e9859c79996b3cc7ab7a8fa6"} Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.176586 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-ltmf9" podStartSLOduration=1.925568138 podStartE2EDuration="3.176558369s" podCreationTimestamp="2026-01-03 03:55:55 +0000 UTC" firstStartedPulling="2026-01-03 03:55:56.298305235 +0000 UTC m=+891.909732059" lastFinishedPulling="2026-01-03 03:55:57.549295456 +0000 UTC m=+893.160722290" observedRunningTime="2026-01-03 03:55:58.170640337 +0000 UTC m=+893.782067191" watchObservedRunningTime="2026-01-03 03:55:58.176558369 +0000 UTC m=+893.787985233" Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.409410 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.523401 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c77z\" (UniqueName: \"kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z\") pod \"162ffb8b-a12f-45bf-84ef-0cdcb1668673\" (UID: \"162ffb8b-a12f-45bf-84ef-0cdcb1668673\") " Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.528344 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z" (OuterVolumeSpecName: "kube-api-access-9c77z") pod "162ffb8b-a12f-45bf-84ef-0cdcb1668673" (UID: "162ffb8b-a12f-45bf-84ef-0cdcb1668673"). InnerVolumeSpecName "kube-api-access-9c77z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:55:58 crc kubenswrapper[4921]: I0103 03:55:58.624412 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c77z\" (UniqueName: \"kubernetes.io/projected/162ffb8b-a12f-45bf-84ef-0cdcb1668673-kube-api-access-9c77z\") on node \"crc\" DevicePath \"\"" Jan 03 03:55:59 crc kubenswrapper[4921]: I0103 03:55:59.162771 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-wdqwl" event={"ID":"162ffb8b-a12f-45bf-84ef-0cdcb1668673","Type":"ContainerDied","Data":"91dcc5660665fe83061fa49fffdc3e920a4f696c24a6faf9bf13b18bdc00ded9"} Jan 03 03:55:59 crc kubenswrapper[4921]: I0103 03:55:59.163125 4921 scope.go:117] "RemoveContainer" containerID="ea70a24eac1f866a1b39acccf9d7a6bd5a325a1e5fe27711a49dd5c68beab50b" Jan 03 03:55:59 crc kubenswrapper[4921]: I0103 03:55:59.162783 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-wdqwl" Jan 03 03:55:59 crc kubenswrapper[4921]: I0103 03:55:59.198508 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:55:59 crc kubenswrapper[4921]: I0103 03:55:59.205397 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-wdqwl"] Jan 03 03:56:00 crc kubenswrapper[4921]: I0103 03:56:00.897360 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" path="/var/lib/kubelet/pods/162ffb8b-a12f-45bf-84ef-0cdcb1668673/volumes" Jan 03 03:56:05 crc kubenswrapper[4921]: I0103 03:56:05.919356 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:56:05 crc kubenswrapper[4921]: I0103 03:56:05.919902 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:56:05 crc kubenswrapper[4921]: I0103 03:56:05.972591 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:56:06 crc kubenswrapper[4921]: I0103 03:56:06.260683 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-ltmf9" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.220610 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr"] Jan 03 03:56:08 crc kubenswrapper[4921]: E0103 03:56:08.221072 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" containerName="registry-server" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.221085 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" containerName="registry-server" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.221180 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="162ffb8b-a12f-45bf-84ef-0cdcb1668673" containerName="registry-server" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.222111 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.226348 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.233060 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr"] Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.378586 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6f75\" (UniqueName: \"kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.378659 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.378693 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.480157 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.480292 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6f75\" (UniqueName: \"kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.480339 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.480917 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.481044 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.517168 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6f75\" (UniqueName: \"kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75\") pod \"676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.571425 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:08 crc kubenswrapper[4921]: I0103 03:56:08.996395 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr"] Jan 03 03:56:09 crc kubenswrapper[4921]: W0103 03:56:09.002224 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod257464fd_1827_4d32_a0bc_b37c100622a0.slice/crio-f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06 WatchSource:0}: Error finding container f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06: Status 404 returned error can't find the container with id f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06 Jan 03 03:56:09 crc kubenswrapper[4921]: I0103 03:56:09.240799 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerStarted","Data":"d442825747eeb686b294d3c114cf9c928c1a0368de29e502433b70920c0e4409"} Jan 03 03:56:09 crc kubenswrapper[4921]: I0103 03:56:09.240855 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerStarted","Data":"f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06"} Jan 03 03:56:10 crc kubenswrapper[4921]: I0103 03:56:10.254109 4921 generic.go:334] "Generic (PLEG): container finished" podID="257464fd-1827-4d32-a0bc-b37c100622a0" containerID="d442825747eeb686b294d3c114cf9c928c1a0368de29e502433b70920c0e4409" exitCode=0 Jan 03 03:56:10 crc kubenswrapper[4921]: I0103 03:56:10.254251 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerDied","Data":"d442825747eeb686b294d3c114cf9c928c1a0368de29e502433b70920c0e4409"} Jan 03 03:56:11 crc kubenswrapper[4921]: I0103 03:56:11.261687 4921 generic.go:334] "Generic (PLEG): container finished" podID="257464fd-1827-4d32-a0bc-b37c100622a0" containerID="2dcf3bae96c9ebd8c19c3173c74135bbd2a9196c6c5001cc3118f127fcb8b394" exitCode=0 Jan 03 03:56:11 crc kubenswrapper[4921]: I0103 03:56:11.261756 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerDied","Data":"2dcf3bae96c9ebd8c19c3173c74135bbd2a9196c6c5001cc3118f127fcb8b394"} Jan 03 03:56:12 crc kubenswrapper[4921]: I0103 03:56:12.271367 4921 generic.go:334] "Generic (PLEG): container finished" podID="257464fd-1827-4d32-a0bc-b37c100622a0" containerID="cb1ce1eb08691e56814482c0090fa897b2f3fefa93940bb944a9bb098e925bb1" exitCode=0 Jan 03 03:56:12 crc kubenswrapper[4921]: I0103 03:56:12.271447 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerDied","Data":"cb1ce1eb08691e56814482c0090fa897b2f3fefa93940bb944a9bb098e925bb1"} Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.611943 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.659331 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util\") pod \"257464fd-1827-4d32-a0bc-b37c100622a0\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.659402 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6f75\" (UniqueName: \"kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75\") pod \"257464fd-1827-4d32-a0bc-b37c100622a0\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.659469 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle\") pod \"257464fd-1827-4d32-a0bc-b37c100622a0\" (UID: \"257464fd-1827-4d32-a0bc-b37c100622a0\") " Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.660594 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle" (OuterVolumeSpecName: "bundle") pod "257464fd-1827-4d32-a0bc-b37c100622a0" (UID: "257464fd-1827-4d32-a0bc-b37c100622a0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.667342 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75" (OuterVolumeSpecName: "kube-api-access-s6f75") pod "257464fd-1827-4d32-a0bc-b37c100622a0" (UID: "257464fd-1827-4d32-a0bc-b37c100622a0"). InnerVolumeSpecName "kube-api-access-s6f75". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.687092 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util" (OuterVolumeSpecName: "util") pod "257464fd-1827-4d32-a0bc-b37c100622a0" (UID: "257464fd-1827-4d32-a0bc-b37c100622a0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.761551 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6f75\" (UniqueName: \"kubernetes.io/projected/257464fd-1827-4d32-a0bc-b37c100622a0-kube-api-access-s6f75\") on node \"crc\" DevicePath \"\"" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.761588 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:56:13 crc kubenswrapper[4921]: I0103 03:56:13.761601 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/257464fd-1827-4d32-a0bc-b37c100622a0-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:56:14 crc kubenswrapper[4921]: I0103 03:56:14.289417 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" event={"ID":"257464fd-1827-4d32-a0bc-b37c100622a0","Type":"ContainerDied","Data":"f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06"} Jan 03 03:56:14 crc kubenswrapper[4921]: I0103 03:56:14.289475 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3212275b397bf1a04610030bb67834cd452aa069b89ac857eff54043f6f4c06" Jan 03 03:56:14 crc kubenswrapper[4921]: I0103 03:56:14.289596 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.113856 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7"] Jan 03 03:56:20 crc kubenswrapper[4921]: E0103 03:56:20.115263 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="util" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.115423 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="util" Jan 03 03:56:20 crc kubenswrapper[4921]: E0103 03:56:20.115481 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="pull" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.115532 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="pull" Jan 03 03:56:20 crc kubenswrapper[4921]: E0103 03:56:20.115592 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="extract" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.115647 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="extract" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.115801 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="257464fd-1827-4d32-a0bc-b37c100622a0" containerName="extract" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.116517 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.121289 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-g89dp" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.121295 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.138994 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7"] Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.159138 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-apiservice-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.159297 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2tz5\" (UniqueName: \"kubernetes.io/projected/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-kube-api-access-j2tz5\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.159328 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-webhook-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.259964 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-apiservice-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.260089 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2tz5\" (UniqueName: \"kubernetes.io/projected/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-kube-api-access-j2tz5\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.260108 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-webhook-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.265379 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-webhook-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.265394 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-apiservice-cert\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.280928 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2tz5\" (UniqueName: \"kubernetes.io/projected/f7d4e479-994e-4d8f-ba48-f5c4961a8d0d-kube-api-access-j2tz5\") pod \"infra-operator-controller-manager-c6c695967-t8mx7\" (UID: \"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d\") " pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.435885 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:20 crc kubenswrapper[4921]: I0103 03:56:20.697413 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7"] Jan 03 03:56:21 crc kubenswrapper[4921]: I0103 03:56:21.338833 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" event={"ID":"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d","Type":"ContainerStarted","Data":"94b61191a48667ad727ce960d3fb018a5577ce264f870b228692fcdb87c33556"} Jan 03 03:56:24 crc kubenswrapper[4921]: I0103 03:56:24.357572 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" event={"ID":"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d","Type":"ContainerStarted","Data":"8b8a22bd2bca3e10bce7c0c288851f518446841b3b47cd2afdcce81cba88545e"} Jan 03 03:56:25 crc kubenswrapper[4921]: I0103 03:56:25.364649 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" event={"ID":"f7d4e479-994e-4d8f-ba48-f5c4961a8d0d","Type":"ContainerStarted","Data":"a95af9a5dee80455d0cceb4f261cd204f538413c3794ad013f987762c13a86af"} Jan 03 03:56:25 crc kubenswrapper[4921]: I0103 03:56:25.365649 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:25 crc kubenswrapper[4921]: I0103 03:56:25.385938 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" podStartSLOduration=1.9434002110000002 podStartE2EDuration="5.38590905s" podCreationTimestamp="2026-01-03 03:56:20 +0000 UTC" firstStartedPulling="2026-01-03 03:56:20.712395985 +0000 UTC m=+916.323822809" lastFinishedPulling="2026-01-03 03:56:24.154904824 +0000 UTC m=+919.766331648" observedRunningTime="2026-01-03 03:56:25.381999673 +0000 UTC m=+920.993426497" watchObservedRunningTime="2026-01-03 03:56:25.38590905 +0000 UTC m=+920.997335874" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.267137 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.268451 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.273902 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-64v2s" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.274328 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.274841 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.274859 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.275050 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.274873 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.292497 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.304640 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.306049 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.323535 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.324825 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.329041 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.346852 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463316 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-default\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463388 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463445 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463474 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463496 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-kolla-config\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463525 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-kolla-config\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463546 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-generated\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463568 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7f8t\" (UniqueName: \"kubernetes.io/projected/615f93ea-16ce-45da-88e6-237f1d57716a-kube-api-access-h7f8t\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463589 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-default\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463612 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.463633 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj9sz\" (UniqueName: \"kubernetes.io/projected/2a9d550e-62e6-4dee-a034-d3e1ea269820-kube-api-access-hj9sz\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466192 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466303 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-default\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466409 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466451 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-operator-scripts\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466480 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-kolla-config\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466505 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5dac2bf9-ef51-4e90-adab-2a73b463967a-secrets\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466537 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lwpm\" (UniqueName: \"kubernetes.io/projected/5dac2bf9-ef51-4e90-adab-2a73b463967a-kube-api-access-5lwpm\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466557 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/615f93ea-16ce-45da-88e6-237f1d57716a-secrets\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466695 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.466754 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2a9d550e-62e6-4dee-a034-d3e1ea269820-secrets\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567646 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567692 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-operator-scripts\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567712 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-kolla-config\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567729 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5dac2bf9-ef51-4e90-adab-2a73b463967a-secrets\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567760 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lwpm\" (UniqueName: \"kubernetes.io/projected/5dac2bf9-ef51-4e90-adab-2a73b463967a-kube-api-access-5lwpm\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567782 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/615f93ea-16ce-45da-88e6-237f1d57716a-secrets\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567808 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567824 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2a9d550e-62e6-4dee-a034-d3e1ea269820-secrets\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567867 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-default\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567899 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567925 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567951 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567974 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-kolla-config\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.567996 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-kolla-config\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568013 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-generated\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568037 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7f8t\" (UniqueName: \"kubernetes.io/projected/615f93ea-16ce-45da-88e6-237f1d57716a-kube-api-access-h7f8t\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568060 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568080 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-default\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568101 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj9sz\" (UniqueName: \"kubernetes.io/projected/2a9d550e-62e6-4dee-a034-d3e1ea269820-kube-api-access-hj9sz\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568130 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.568151 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-default\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.569156 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-default\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.569162 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.569237 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-kolla-config\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.569252 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.569325 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.570383 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-default\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.570671 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-generated\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.570718 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-kolla-config\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.571212 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dac2bf9-ef51-4e90-adab-2a73b463967a-operator-scripts\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.571427 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/615f93ea-16ce-45da-88e6-237f1d57716a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.571721 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.572020 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/615f93ea-16ce-45da-88e6-237f1d57716a-kolla-config\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.573617 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-config-data-default\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.581048 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/615f93ea-16ce-45da-88e6-237f1d57716a-secrets\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.581556 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5dac2bf9-ef51-4e90-adab-2a73b463967a-secrets\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.582073 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2a9d550e-62e6-4dee-a034-d3e1ea269820-secrets\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.584556 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a9d550e-62e6-4dee-a034-d3e1ea269820-operator-scripts\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.591398 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj9sz\" (UniqueName: \"kubernetes.io/projected/2a9d550e-62e6-4dee-a034-d3e1ea269820-kube-api-access-hj9sz\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.598662 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5dac2bf9-ef51-4e90-adab-2a73b463967a-config-data-generated\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.600169 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7f8t\" (UniqueName: \"kubernetes.io/projected/615f93ea-16ce-45da-88e6-237f1d57716a-kube-api-access-h7f8t\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.601337 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"615f93ea-16ce-45da-88e6-237f1d57716a\") " pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.602425 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lwpm\" (UniqueName: \"kubernetes.io/projected/5dac2bf9-ef51-4e90-adab-2a73b463967a-kube-api-access-5lwpm\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.611468 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"2a9d550e-62e6-4dee-a034-d3e1ea269820\") " pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.617941 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-2\" (UID: \"5dac2bf9-ef51-4e90-adab-2a73b463967a\") " pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.628667 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.641337 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:26 crc kubenswrapper[4921]: I0103 03:56:26.891826 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.052837 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Jan 03 03:56:27 crc kubenswrapper[4921]: W0103 03:56:27.062751 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dac2bf9_ef51_4e90_adab_2a73b463967a.slice/crio-8a88d45862313d7145412fa4432752d53c98516fde3a8a787b84a451d41c42dd WatchSource:0}: Error finding container 8a88d45862313d7145412fa4432752d53c98516fde3a8a787b84a451d41c42dd: Status 404 returned error can't find the container with id 8a88d45862313d7145412fa4432752d53c98516fde3a8a787b84a451d41c42dd Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.114708 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Jan 03 03:56:27 crc kubenswrapper[4921]: W0103 03:56:27.121864 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a9d550e_62e6_4dee_a034_d3e1ea269820.slice/crio-b4c5df669c5fdc825bdc4bdbdf9cf1a3148c45dccaf9fce2dcfc930bf9e072e2 WatchSource:0}: Error finding container b4c5df669c5fdc825bdc4bdbdf9cf1a3148c45dccaf9fce2dcfc930bf9e072e2: Status 404 returned error can't find the container with id b4c5df669c5fdc825bdc4bdbdf9cf1a3148c45dccaf9fce2dcfc930bf9e072e2 Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.331476 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Jan 03 03:56:27 crc kubenswrapper[4921]: W0103 03:56:27.337924 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod615f93ea_16ce_45da_88e6_237f1d57716a.slice/crio-141e63509d55fea4ea7e16e76ab204e3b12f84a2c17ea00fa4b3f95202649a9d WatchSource:0}: Error finding container 141e63509d55fea4ea7e16e76ab204e3b12f84a2c17ea00fa4b3f95202649a9d: Status 404 returned error can't find the container with id 141e63509d55fea4ea7e16e76ab204e3b12f84a2c17ea00fa4b3f95202649a9d Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.386750 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"2a9d550e-62e6-4dee-a034-d3e1ea269820","Type":"ContainerStarted","Data":"b4c5df669c5fdc825bdc4bdbdf9cf1a3148c45dccaf9fce2dcfc930bf9e072e2"} Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.388358 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"5dac2bf9-ef51-4e90-adab-2a73b463967a","Type":"ContainerStarted","Data":"8a88d45862313d7145412fa4432752d53c98516fde3a8a787b84a451d41c42dd"} Jan 03 03:56:27 crc kubenswrapper[4921]: I0103 03:56:27.389772 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"615f93ea-16ce-45da-88e6-237f1d57716a","Type":"ContainerStarted","Data":"141e63509d55fea4ea7e16e76ab204e3b12f84a2c17ea00fa4b3f95202649a9d"} Jan 03 03:56:30 crc kubenswrapper[4921]: I0103 03:56:30.454131 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-c6c695967-t8mx7" Jan 03 03:56:31 crc kubenswrapper[4921]: I0103 03:56:31.175372 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:56:31 crc kubenswrapper[4921]: I0103 03:56:31.175661 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:56:35 crc kubenswrapper[4921]: I0103 03:56:35.457941 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"615f93ea-16ce-45da-88e6-237f1d57716a","Type":"ContainerStarted","Data":"fc614e8be16751a105e79fd79ac94de0196d5e40ae0e74e7897338a126a3f1c1"} Jan 03 03:56:35 crc kubenswrapper[4921]: I0103 03:56:35.461908 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"2a9d550e-62e6-4dee-a034-d3e1ea269820","Type":"ContainerStarted","Data":"ddbfe600c28254a6c218d61b4da2470b975f90fb8203082d72e95f8933055e65"} Jan 03 03:56:35 crc kubenswrapper[4921]: I0103 03:56:35.463376 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"5dac2bf9-ef51-4e90-adab-2a73b463967a","Type":"ContainerStarted","Data":"2787e2bed9acc3fc6103975a975f6c2942a3be866f7aa1970a29f1cd4564bdca"} Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.181042 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-277lk"] Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.182227 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.184952 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-zhsk6" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.190656 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-277lk"] Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.244527 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-685ld\" (UniqueName: \"kubernetes.io/projected/69284c8d-7396-4067-adad-72b4e024b197-kube-api-access-685ld\") pod \"rabbitmq-cluster-operator-index-277lk\" (UID: \"69284c8d-7396-4067-adad-72b4e024b197\") " pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.346222 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-685ld\" (UniqueName: \"kubernetes.io/projected/69284c8d-7396-4067-adad-72b4e024b197-kube-api-access-685ld\") pod \"rabbitmq-cluster-operator-index-277lk\" (UID: \"69284c8d-7396-4067-adad-72b4e024b197\") " pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.364210 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-685ld\" (UniqueName: \"kubernetes.io/projected/69284c8d-7396-4067-adad-72b4e024b197-kube-api-access-685ld\") pod \"rabbitmq-cluster-operator-index-277lk\" (UID: \"69284c8d-7396-4067-adad-72b4e024b197\") " pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.499316 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:37 crc kubenswrapper[4921]: I0103 03:56:37.966733 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-277lk"] Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.486694 4921 generic.go:334] "Generic (PLEG): container finished" podID="2a9d550e-62e6-4dee-a034-d3e1ea269820" containerID="ddbfe600c28254a6c218d61b4da2470b975f90fb8203082d72e95f8933055e65" exitCode=0 Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.486830 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"2a9d550e-62e6-4dee-a034-d3e1ea269820","Type":"ContainerDied","Data":"ddbfe600c28254a6c218d61b4da2470b975f90fb8203082d72e95f8933055e65"} Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.489724 4921 generic.go:334] "Generic (PLEG): container finished" podID="5dac2bf9-ef51-4e90-adab-2a73b463967a" containerID="2787e2bed9acc3fc6103975a975f6c2942a3be866f7aa1970a29f1cd4564bdca" exitCode=0 Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.489819 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"5dac2bf9-ef51-4e90-adab-2a73b463967a","Type":"ContainerDied","Data":"2787e2bed9acc3fc6103975a975f6c2942a3be866f7aa1970a29f1cd4564bdca"} Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.491637 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" event={"ID":"69284c8d-7396-4067-adad-72b4e024b197","Type":"ContainerStarted","Data":"429c06750c73c7f0521107f9100e9b6421d59118b7cfc18e1fce48bf96289ada"} Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.493740 4921 generic.go:334] "Generic (PLEG): container finished" podID="615f93ea-16ce-45da-88e6-237f1d57716a" containerID="fc614e8be16751a105e79fd79ac94de0196d5e40ae0e74e7897338a126a3f1c1" exitCode=0 Jan 03 03:56:38 crc kubenswrapper[4921]: I0103 03:56:38.493785 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"615f93ea-16ce-45da-88e6-237f1d57716a","Type":"ContainerDied","Data":"fc614e8be16751a105e79fd79ac94de0196d5e40ae0e74e7897338a126a3f1c1"} Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.504624 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"615f93ea-16ce-45da-88e6-237f1d57716a","Type":"ContainerStarted","Data":"84ef32ebcf1d147d634f1b4eb5ecba3f501bee86c6db02464d8c41389059aba2"} Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.507047 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"2a9d550e-62e6-4dee-a034-d3e1ea269820","Type":"ContainerStarted","Data":"3354ace0f0336b2f347369ed9d949f307e1a23c89c564d4fe2f91ddf2597c359"} Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.509034 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"5dac2bf9-ef51-4e90-adab-2a73b463967a","Type":"ContainerStarted","Data":"2d7010e51f21132ade94c5e0297eee15407f4808b1a823ac48796ab669210c1e"} Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.533133 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=7.532956013 podStartE2EDuration="14.533113215s" podCreationTimestamp="2026-01-03 03:56:25 +0000 UTC" firstStartedPulling="2026-01-03 03:56:27.339987813 +0000 UTC m=+922.951414657" lastFinishedPulling="2026-01-03 03:56:34.340145035 +0000 UTC m=+929.951571859" observedRunningTime="2026-01-03 03:56:39.525771985 +0000 UTC m=+935.137198829" watchObservedRunningTime="2026-01-03 03:56:39.533113215 +0000 UTC m=+935.144540049" Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.550193 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=7.275130572 podStartE2EDuration="14.550178841s" podCreationTimestamp="2026-01-03 03:56:25 +0000 UTC" firstStartedPulling="2026-01-03 03:56:27.066404831 +0000 UTC m=+922.677831655" lastFinishedPulling="2026-01-03 03:56:34.34145309 +0000 UTC m=+929.952879924" observedRunningTime="2026-01-03 03:56:39.545871823 +0000 UTC m=+935.157298667" watchObservedRunningTime="2026-01-03 03:56:39.550178841 +0000 UTC m=+935.161605665" Jan 03 03:56:39 crc kubenswrapper[4921]: I0103 03:56:39.575370 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=7.392697813 podStartE2EDuration="14.575350468s" podCreationTimestamp="2026-01-03 03:56:25 +0000 UTC" firstStartedPulling="2026-01-03 03:56:27.124669123 +0000 UTC m=+922.736095947" lastFinishedPulling="2026-01-03 03:56:34.307321748 +0000 UTC m=+929.918748602" observedRunningTime="2026-01-03 03:56:39.565966412 +0000 UTC m=+935.177393256" watchObservedRunningTime="2026-01-03 03:56:39.575350468 +0000 UTC m=+935.186777292" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.583449 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.584728 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.588226 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.588322 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-xqrmv" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.597450 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.779283 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kolla-config\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.779329 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4dvk\" (UniqueName: \"kubernetes.io/projected/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kube-api-access-l4dvk\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.779361 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-config-data\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.880481 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kolla-config\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.880822 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4dvk\" (UniqueName: \"kubernetes.io/projected/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kube-api-access-l4dvk\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.880865 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-config-data\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.881819 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-config-data\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.882061 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kolla-config\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.902544 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4dvk\" (UniqueName: \"kubernetes.io/projected/67fd2f86-99b0-4dfa-a40e-f3438940fa71-kube-api-access-l4dvk\") pod \"memcached-0\" (UID: \"67fd2f86-99b0-4dfa-a40e-f3438940fa71\") " pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:44 crc kubenswrapper[4921]: I0103 03:56:44.902855 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:45 crc kubenswrapper[4921]: I0103 03:56:45.338803 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Jan 03 03:56:45 crc kubenswrapper[4921]: I0103 03:56:45.580616 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"67fd2f86-99b0-4dfa-a40e-f3438940fa71","Type":"ContainerStarted","Data":"900961528ea3b44ea67eae7c709fa4ec420abe3040b640f18f8595054b55a52e"} Jan 03 03:56:45 crc kubenswrapper[4921]: I0103 03:56:45.581979 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" event={"ID":"69284c8d-7396-4067-adad-72b4e024b197","Type":"ContainerStarted","Data":"759f5b2829b5126257953c6bd8abadc384771632b48eac136dccff9dac714061"} Jan 03 03:56:45 crc kubenswrapper[4921]: I0103 03:56:45.603773 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" podStartSLOduration=2.14332964 podStartE2EDuration="8.603753843s" podCreationTimestamp="2026-01-03 03:56:37 +0000 UTC" firstStartedPulling="2026-01-03 03:56:37.975672904 +0000 UTC m=+933.587099728" lastFinishedPulling="2026-01-03 03:56:44.436097067 +0000 UTC m=+940.047523931" observedRunningTime="2026-01-03 03:56:45.596744032 +0000 UTC m=+941.208170906" watchObservedRunningTime="2026-01-03 03:56:45.603753843 +0000 UTC m=+941.215180687" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.629456 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.629523 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.641674 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.641714 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.900889 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:46 crc kubenswrapper[4921]: I0103 03:56:46.900936 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:56:47 crc kubenswrapper[4921]: I0103 03:56:47.499744 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:47 crc kubenswrapper[4921]: I0103 03:56:47.499832 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:47 crc kubenswrapper[4921]: I0103 03:56:47.535209 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:50 crc kubenswrapper[4921]: I0103 03:56:50.687918 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:50 crc kubenswrapper[4921]: I0103 03:56:50.761065 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Jan 03 03:56:51 crc kubenswrapper[4921]: I0103 03:56:51.621179 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"67fd2f86-99b0-4dfa-a40e-f3438940fa71","Type":"ContainerStarted","Data":"c5fb9de1d7e19c12100fd4262909871995a26f3d891b35a24aafb5e2c6158959"} Jan 03 03:56:51 crc kubenswrapper[4921]: I0103 03:56:51.621701 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Jan 03 03:56:51 crc kubenswrapper[4921]: I0103 03:56:51.643985 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=2.011723455 podStartE2EDuration="7.643968922s" podCreationTimestamp="2026-01-03 03:56:44 +0000 UTC" firstStartedPulling="2026-01-03 03:56:45.343501906 +0000 UTC m=+940.954928750" lastFinishedPulling="2026-01-03 03:56:50.975747393 +0000 UTC m=+946.587174217" observedRunningTime="2026-01-03 03:56:51.64135082 +0000 UTC m=+947.252777644" watchObservedRunningTime="2026-01-03 03:56:51.643968922 +0000 UTC m=+947.255395746" Jan 03 03:56:56 crc kubenswrapper[4921]: I0103 03:56:56.682934 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="5dac2bf9-ef51-4e90-adab-2a73b463967a" containerName="galera" probeResult="failure" output=< Jan 03 03:56:56 crc kubenswrapper[4921]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Jan 03 03:56:56 crc kubenswrapper[4921]: > Jan 03 03:56:57 crc kubenswrapper[4921]: I0103 03:56:57.543594 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-277lk" Jan 03 03:56:59 crc kubenswrapper[4921]: I0103 03:56:59.904142 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Jan 03 03:57:01 crc kubenswrapper[4921]: I0103 03:57:01.176100 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:57:01 crc kubenswrapper[4921]: I0103 03:57:01.176194 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:57:04 crc kubenswrapper[4921]: I0103 03:57:04.667300 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:57:04 crc kubenswrapper[4921]: I0103 03:57:04.735566 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Jan 03 03:57:05 crc kubenswrapper[4921]: I0103 03:57:05.506111 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:57:05 crc kubenswrapper[4921]: I0103 03:57:05.563683 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.018407 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf"] Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.019719 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.021418 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.031818 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf"] Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.108896 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.109145 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhk78\" (UniqueName: \"kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.109262 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.211007 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.211135 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.211178 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhk78\" (UniqueName: \"kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.211827 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.211916 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.233627 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhk78\" (UniqueName: \"kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.341559 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.567131 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf"] Jan 03 03:57:06 crc kubenswrapper[4921]: W0103 03:57:06.577533 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23d84029_73ca_4284_94e4_717bb7c8463d.slice/crio-f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783 WatchSource:0}: Error finding container f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783: Status 404 returned error can't find the container with id f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783 Jan 03 03:57:06 crc kubenswrapper[4921]: I0103 03:57:06.721483 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" event={"ID":"23d84029-73ca-4284-94e4-717bb7c8463d","Type":"ContainerStarted","Data":"f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783"} Jan 03 03:57:07 crc kubenswrapper[4921]: I0103 03:57:07.730816 4921 generic.go:334] "Generic (PLEG): container finished" podID="23d84029-73ca-4284-94e4-717bb7c8463d" containerID="25f012b646b4f59a49d72b62ea318542a00d72cfb0cdb7c55f149d455c31a873" exitCode=0 Jan 03 03:57:07 crc kubenswrapper[4921]: I0103 03:57:07.730872 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" event={"ID":"23d84029-73ca-4284-94e4-717bb7c8463d","Type":"ContainerDied","Data":"25f012b646b4f59a49d72b62ea318542a00d72cfb0cdb7c55f149d455c31a873"} Jan 03 03:57:08 crc kubenswrapper[4921]: I0103 03:57:08.743164 4921 generic.go:334] "Generic (PLEG): container finished" podID="23d84029-73ca-4284-94e4-717bb7c8463d" containerID="51aa9eaa8dabbf3df6da728c8823e5189996692002155e63b29ee93d7aca90bb" exitCode=0 Jan 03 03:57:08 crc kubenswrapper[4921]: I0103 03:57:08.743258 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" event={"ID":"23d84029-73ca-4284-94e4-717bb7c8463d","Type":"ContainerDied","Data":"51aa9eaa8dabbf3df6da728c8823e5189996692002155e63b29ee93d7aca90bb"} Jan 03 03:57:09 crc kubenswrapper[4921]: I0103 03:57:09.756372 4921 generic.go:334] "Generic (PLEG): container finished" podID="23d84029-73ca-4284-94e4-717bb7c8463d" containerID="a285c511f46d6e5b13591985ce2661eb112e3a6839ee57016edb39ab94037961" exitCode=0 Jan 03 03:57:09 crc kubenswrapper[4921]: I0103 03:57:09.756437 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" event={"ID":"23d84029-73ca-4284-94e4-717bb7c8463d","Type":"ContainerDied","Data":"a285c511f46d6e5b13591985ce2661eb112e3a6839ee57016edb39ab94037961"} Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.086760 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.100798 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle\") pod \"23d84029-73ca-4284-94e4-717bb7c8463d\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.100936 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util\") pod \"23d84029-73ca-4284-94e4-717bb7c8463d\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.101158 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhk78\" (UniqueName: \"kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78\") pod \"23d84029-73ca-4284-94e4-717bb7c8463d\" (UID: \"23d84029-73ca-4284-94e4-717bb7c8463d\") " Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.102316 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle" (OuterVolumeSpecName: "bundle") pod "23d84029-73ca-4284-94e4-717bb7c8463d" (UID: "23d84029-73ca-4284-94e4-717bb7c8463d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.102772 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.115631 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78" (OuterVolumeSpecName: "kube-api-access-vhk78") pod "23d84029-73ca-4284-94e4-717bb7c8463d" (UID: "23d84029-73ca-4284-94e4-717bb7c8463d"). InnerVolumeSpecName "kube-api-access-vhk78". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.127037 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util" (OuterVolumeSpecName: "util") pod "23d84029-73ca-4284-94e4-717bb7c8463d" (UID: "23d84029-73ca-4284-94e4-717bb7c8463d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.203968 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/23d84029-73ca-4284-94e4-717bb7c8463d-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.204010 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhk78\" (UniqueName: \"kubernetes.io/projected/23d84029-73ca-4284-94e4-717bb7c8463d-kube-api-access-vhk78\") on node \"crc\" DevicePath \"\"" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.790380 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" event={"ID":"23d84029-73ca-4284-94e4-717bb7c8463d","Type":"ContainerDied","Data":"f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783"} Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.791069 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3f60f5b9486760a7a035c69ae4718676ac7912ad83b92d4d585efd4b0232783" Jan 03 03:57:11 crc kubenswrapper[4921]: I0103 03:57:11.790520 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.891574 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p"] Jan 03 03:57:19 crc kubenswrapper[4921]: E0103 03:57:19.892258 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="extract" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.892290 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="extract" Jan 03 03:57:19 crc kubenswrapper[4921]: E0103 03:57:19.892308 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="pull" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.892317 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="pull" Jan 03 03:57:19 crc kubenswrapper[4921]: E0103 03:57:19.892339 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="util" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.892348 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="util" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.892468 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d84029-73ca-4284-94e4-717bb7c8463d" containerName="extract" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.892931 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.895089 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-dz55x" Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.910783 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p"] Jan 03 03:57:19 crc kubenswrapper[4921]: I0103 03:57:19.930337 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbqgn\" (UniqueName: \"kubernetes.io/projected/bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf-kube-api-access-cbqgn\") pod \"rabbitmq-cluster-operator-779fc9694b-hd77p\" (UID: \"bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" Jan 03 03:57:20 crc kubenswrapper[4921]: I0103 03:57:20.032720 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbqgn\" (UniqueName: \"kubernetes.io/projected/bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf-kube-api-access-cbqgn\") pod \"rabbitmq-cluster-operator-779fc9694b-hd77p\" (UID: \"bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" Jan 03 03:57:20 crc kubenswrapper[4921]: I0103 03:57:20.052241 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbqgn\" (UniqueName: \"kubernetes.io/projected/bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf-kube-api-access-cbqgn\") pod \"rabbitmq-cluster-operator-779fc9694b-hd77p\" (UID: \"bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" Jan 03 03:57:20 crc kubenswrapper[4921]: I0103 03:57:20.223890 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" Jan 03 03:57:20 crc kubenswrapper[4921]: I0103 03:57:20.728432 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p"] Jan 03 03:57:20 crc kubenswrapper[4921]: I0103 03:57:20.857261 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" event={"ID":"bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf","Type":"ContainerStarted","Data":"dca01866a82927b90d104580b0c767b2e8cfd05750d3839c71946e113bf1109c"} Jan 03 03:57:23 crc kubenswrapper[4921]: I0103 03:57:23.877382 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" event={"ID":"bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf","Type":"ContainerStarted","Data":"f29864e3bbb7e5a389c9a7eb9b6188e29f7c5e7fff8a1a34784edb8d2c85813b"} Jan 03 03:57:23 crc kubenswrapper[4921]: I0103 03:57:23.897528 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-hd77p" podStartSLOduration=2.097663853 podStartE2EDuration="4.897507659s" podCreationTimestamp="2026-01-03 03:57:19 +0000 UTC" firstStartedPulling="2026-01-03 03:57:20.74894391 +0000 UTC m=+976.360370734" lastFinishedPulling="2026-01-03 03:57:23.548787716 +0000 UTC m=+979.160214540" observedRunningTime="2026-01-03 03:57:23.896909312 +0000 UTC m=+979.508336196" watchObservedRunningTime="2026-01-03 03:57:23.897507659 +0000 UTC m=+979.508934493" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.158909 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.160542 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.163047 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-8fcjq" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.163234 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.163246 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.164072 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.164666 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.177305 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178091 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178157 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26tzr\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-kube-api-access-26tzr\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178239 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178297 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178363 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178396 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178432 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.178463 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279605 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279671 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279712 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279746 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279822 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279859 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26tzr\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-kube-api-access-26tzr\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279929 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.279967 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.280176 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.280557 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.281174 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.285108 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.285779 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.285867 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.286099 4921 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.286125 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/3c98394e30b941f306e9154d770dbc474cd4e70019371a34c5becc87f133fbd5/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.313411 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26tzr\" (UniqueName: \"kubernetes.io/projected/e3b45f46-86ca-4ffb-b451-00d8631b9aaa-kube-api-access-26tzr\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.321720 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b13fe710-455c-4ee1-a0c4-7490e9792002\") pod \"rabbitmq-server-0\" (UID: \"e3b45f46-86ca-4ffb-b451-00d8631b9aaa\") " pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.480160 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.727344 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Jan 03 03:57:27 crc kubenswrapper[4921]: I0103 03:57:27.902878 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"e3b45f46-86ca-4ffb-b451-00d8631b9aaa","Type":"ContainerStarted","Data":"7c61fde7b011814021748f970c9c7e11e8cd1678955b614f71173565f92a8520"} Jan 03 03:57:28 crc kubenswrapper[4921]: I0103 03:57:28.791530 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:28 crc kubenswrapper[4921]: I0103 03:57:28.792619 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:28 crc kubenswrapper[4921]: I0103 03:57:28.795385 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-rjwzp" Jan 03 03:57:28 crc kubenswrapper[4921]: I0103 03:57:28.798650 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:28 crc kubenswrapper[4921]: I0103 03:57:28.898674 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr665\" (UniqueName: \"kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665\") pod \"keystone-operator-index-hq67b\" (UID: \"d4a96509-e3bc-493a-8569-3f57a5d51945\") " pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:29 crc kubenswrapper[4921]: I0103 03:57:29.000143 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr665\" (UniqueName: \"kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665\") pod \"keystone-operator-index-hq67b\" (UID: \"d4a96509-e3bc-493a-8569-3f57a5d51945\") " pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:29 crc kubenswrapper[4921]: I0103 03:57:29.019923 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr665\" (UniqueName: \"kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665\") pod \"keystone-operator-index-hq67b\" (UID: \"d4a96509-e3bc-493a-8569-3f57a5d51945\") " pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:29 crc kubenswrapper[4921]: I0103 03:57:29.114358 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:29 crc kubenswrapper[4921]: I0103 03:57:29.522145 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:29 crc kubenswrapper[4921]: W0103 03:57:29.526452 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4a96509_e3bc_493a_8569_3f57a5d51945.slice/crio-60f37a35d4cc79801ff19973cd6aa0c993ac676370d0127155a95f3a40be136e WatchSource:0}: Error finding container 60f37a35d4cc79801ff19973cd6aa0c993ac676370d0127155a95f3a40be136e: Status 404 returned error can't find the container with id 60f37a35d4cc79801ff19973cd6aa0c993ac676370d0127155a95f3a40be136e Jan 03 03:57:29 crc kubenswrapper[4921]: I0103 03:57:29.926573 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hq67b" event={"ID":"d4a96509-e3bc-493a-8569-3f57a5d51945","Type":"ContainerStarted","Data":"60f37a35d4cc79801ff19973cd6aa0c993ac676370d0127155a95f3a40be136e"} Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.175538 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.175903 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.175954 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.176649 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.176725 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd" gracePeriod=600 Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.946812 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd" exitCode=0 Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.946897 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd"} Jan 03 03:57:31 crc kubenswrapper[4921]: I0103 03:57:31.947010 4921 scope.go:117] "RemoveContainer" containerID="155208290f027b815f00500c17737b19b77dfe5792ee69db7668686c3b2f65f2" Jan 03 03:57:33 crc kubenswrapper[4921]: I0103 03:57:33.179883 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:33 crc kubenswrapper[4921]: I0103 03:57:33.781046 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-fjmv9"] Jan 03 03:57:33 crc kubenswrapper[4921]: I0103 03:57:33.781946 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:33 crc kubenswrapper[4921]: I0103 03:57:33.797342 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-fjmv9"] Jan 03 03:57:34 crc kubenswrapper[4921]: I0103 03:57:34.014386 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckkfd\" (UniqueName: \"kubernetes.io/projected/2486f9b9-492d-47e6-9620-9a1b53d419eb-kube-api-access-ckkfd\") pod \"keystone-operator-index-fjmv9\" (UID: \"2486f9b9-492d-47e6-9620-9a1b53d419eb\") " pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:34 crc kubenswrapper[4921]: I0103 03:57:34.118385 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckkfd\" (UniqueName: \"kubernetes.io/projected/2486f9b9-492d-47e6-9620-9a1b53d419eb-kube-api-access-ckkfd\") pod \"keystone-operator-index-fjmv9\" (UID: \"2486f9b9-492d-47e6-9620-9a1b53d419eb\") " pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:34 crc kubenswrapper[4921]: I0103 03:57:34.146985 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckkfd\" (UniqueName: \"kubernetes.io/projected/2486f9b9-492d-47e6-9620-9a1b53d419eb-kube-api-access-ckkfd\") pod \"keystone-operator-index-fjmv9\" (UID: \"2486f9b9-492d-47e6-9620-9a1b53d419eb\") " pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:34 crc kubenswrapper[4921]: I0103 03:57:34.223895 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:35 crc kubenswrapper[4921]: I0103 03:57:35.456370 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-fjmv9"] Jan 03 03:57:35 crc kubenswrapper[4921]: W0103 03:57:35.822048 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2486f9b9_492d_47e6_9620_9a1b53d419eb.slice/crio-2a1df66897704316cf73e3a78851c3864ee561750cfd1885e4de0c2f6e98181e WatchSource:0}: Error finding container 2a1df66897704316cf73e3a78851c3864ee561750cfd1885e4de0c2f6e98181e: Status 404 returned error can't find the container with id 2a1df66897704316cf73e3a78851c3864ee561750cfd1885e4de0c2f6e98181e Jan 03 03:57:35 crc kubenswrapper[4921]: I0103 03:57:35.977379 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-fjmv9" event={"ID":"2486f9b9-492d-47e6-9620-9a1b53d419eb","Type":"ContainerStarted","Data":"2a1df66897704316cf73e3a78851c3864ee561750cfd1885e4de0c2f6e98181e"} Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.036705 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hq67b" event={"ID":"d4a96509-e3bc-493a-8569-3f57a5d51945","Type":"ContainerStarted","Data":"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d"} Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.036923 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-hq67b" podUID="d4a96509-e3bc-493a-8569-3f57a5d51945" containerName="registry-server" containerID="cri-o://f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d" gracePeriod=2 Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.041433 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-fjmv9" event={"ID":"2486f9b9-492d-47e6-9620-9a1b53d419eb","Type":"ContainerStarted","Data":"040bf9094f539142e268f1a3edcc28cf6cb43760b9f7f3f17fc24289dec0fb5d"} Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.051428 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4"} Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.060546 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-hq67b" podStartSLOduration=2.145364035 podStartE2EDuration="16.060528436s" podCreationTimestamp="2026-01-03 03:57:28 +0000 UTC" firstStartedPulling="2026-01-03 03:57:29.528532472 +0000 UTC m=+985.139959286" lastFinishedPulling="2026-01-03 03:57:43.443696863 +0000 UTC m=+999.055123687" observedRunningTime="2026-01-03 03:57:44.056148182 +0000 UTC m=+999.667575046" watchObservedRunningTime="2026-01-03 03:57:44.060528436 +0000 UTC m=+999.671955270" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.074925 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-fjmv9" podStartSLOduration=3.315487264 podStartE2EDuration="11.074911156s" podCreationTimestamp="2026-01-03 03:57:33 +0000 UTC" firstStartedPulling="2026-01-03 03:57:35.824746453 +0000 UTC m=+991.436173277" lastFinishedPulling="2026-01-03 03:57:43.584170345 +0000 UTC m=+999.195597169" observedRunningTime="2026-01-03 03:57:44.072548668 +0000 UTC m=+999.683975522" watchObservedRunningTime="2026-01-03 03:57:44.074911156 +0000 UTC m=+999.686337980" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.226046 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.226423 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.262607 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.505848 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.685542 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr665\" (UniqueName: \"kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665\") pod \"d4a96509-e3bc-493a-8569-3f57a5d51945\" (UID: \"d4a96509-e3bc-493a-8569-3f57a5d51945\") " Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.695146 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665" (OuterVolumeSpecName: "kube-api-access-nr665") pod "d4a96509-e3bc-493a-8569-3f57a5d51945" (UID: "d4a96509-e3bc-493a-8569-3f57a5d51945"). InnerVolumeSpecName "kube-api-access-nr665". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:57:44 crc kubenswrapper[4921]: I0103 03:57:44.787804 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr665\" (UniqueName: \"kubernetes.io/projected/d4a96509-e3bc-493a-8569-3f57a5d51945-kube-api-access-nr665\") on node \"crc\" DevicePath \"\"" Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.059414 4921 generic.go:334] "Generic (PLEG): container finished" podID="d4a96509-e3bc-493a-8569-3f57a5d51945" containerID="f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d" exitCode=0 Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.059466 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-hq67b" Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.059476 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hq67b" event={"ID":"d4a96509-e3bc-493a-8569-3f57a5d51945","Type":"ContainerDied","Data":"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d"} Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.059501 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-hq67b" event={"ID":"d4a96509-e3bc-493a-8569-3f57a5d51945","Type":"ContainerDied","Data":"60f37a35d4cc79801ff19973cd6aa0c993ac676370d0127155a95f3a40be136e"} Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.059518 4921 scope.go:117] "RemoveContainer" containerID="f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d" Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.064791 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"e3b45f46-86ca-4ffb-b451-00d8631b9aaa","Type":"ContainerStarted","Data":"bc1f1683c7e0e98827171a6cb4b13587562267bfccbe4b7c822e7adc5de9b78c"} Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.086071 4921 scope.go:117] "RemoveContainer" containerID="f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d" Jan 03 03:57:45 crc kubenswrapper[4921]: E0103 03:57:45.086702 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d\": container with ID starting with f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d not found: ID does not exist" containerID="f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d" Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.086776 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d"} err="failed to get container status \"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d\": rpc error: code = NotFound desc = could not find container \"f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d\": container with ID starting with f0879cd4b52f6f9513775da918e3b8a996644606f3977992cf23ff693ab5ab7d not found: ID does not exist" Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.112829 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:45 crc kubenswrapper[4921]: I0103 03:57:45.119009 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-hq67b"] Jan 03 03:57:46 crc kubenswrapper[4921]: I0103 03:57:46.892487 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a96509-e3bc-493a-8569-3f57a5d51945" path="/var/lib/kubelet/pods/d4a96509-e3bc-493a-8569-3f57a5d51945/volumes" Jan 03 03:57:54 crc kubenswrapper[4921]: I0103 03:57:54.268373 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-fjmv9" Jan 03 03:58:04 crc kubenswrapper[4921]: I0103 03:58:04.985573 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:04 crc kubenswrapper[4921]: E0103 03:58:04.987236 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a96509-e3bc-493a-8569-3f57a5d51945" containerName="registry-server" Jan 03 03:58:04 crc kubenswrapper[4921]: I0103 03:58:04.987347 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a96509-e3bc-493a-8569-3f57a5d51945" containerName="registry-server" Jan 03 03:58:04 crc kubenswrapper[4921]: I0103 03:58:04.987548 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a96509-e3bc-493a-8569-3f57a5d51945" containerName="registry-server" Jan 03 03:58:04 crc kubenswrapper[4921]: I0103 03:58:04.988524 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.005739 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.032506 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.032575 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp6zh\" (UniqueName: \"kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.032748 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.133785 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.134102 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.134223 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp6zh\" (UniqueName: \"kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.134333 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.134836 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.157459 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp6zh\" (UniqueName: \"kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh\") pod \"redhat-operators-xzbrh\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.318507 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.774067 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.837832 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k"] Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.838983 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.841891 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.853961 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.854025 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.854149 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h44h\" (UniqueName: \"kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.854954 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k"] Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.955210 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.955614 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h44h\" (UniqueName: \"kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.955714 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.955866 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.956110 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:05 crc kubenswrapper[4921]: I0103 03:58:05.975081 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h44h\" (UniqueName: \"kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h\") pod \"a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:06 crc kubenswrapper[4921]: I0103 03:58:06.184464 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:06 crc kubenswrapper[4921]: I0103 03:58:06.213976 4921 generic.go:334] "Generic (PLEG): container finished" podID="08a322d7-503b-498f-99cb-149fd7032129" containerID="714c78e77d9f76fb74a16602b38cde0b0d9788209f65ceac702214f2bc2e3507" exitCode=0 Jan 03 03:58:06 crc kubenswrapper[4921]: I0103 03:58:06.214022 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerDied","Data":"714c78e77d9f76fb74a16602b38cde0b0d9788209f65ceac702214f2bc2e3507"} Jan 03 03:58:06 crc kubenswrapper[4921]: I0103 03:58:06.214051 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerStarted","Data":"5306f88ddf819c345c2945ab8d14df792cf5a0c7113dd436b42742716d95343b"} Jan 03 03:58:06 crc kubenswrapper[4921]: I0103 03:58:06.405876 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k"] Jan 03 03:58:07 crc kubenswrapper[4921]: I0103 03:58:07.225096 4921 generic.go:334] "Generic (PLEG): container finished" podID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerID="023a361876b05e9419954562293c4646efc62026f5dbe0cada9de58e997078f9" exitCode=0 Jan 03 03:58:07 crc kubenswrapper[4921]: I0103 03:58:07.225182 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" event={"ID":"4ca0e16f-8e0c-455a-9767-951056bd29cb","Type":"ContainerDied","Data":"023a361876b05e9419954562293c4646efc62026f5dbe0cada9de58e997078f9"} Jan 03 03:58:07 crc kubenswrapper[4921]: I0103 03:58:07.225458 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" event={"ID":"4ca0e16f-8e0c-455a-9767-951056bd29cb","Type":"ContainerStarted","Data":"f03340c269b4a0b18c48db63b1dc35624e4766c8cd4cf7faa21b1ecc925633a8"} Jan 03 03:58:08 crc kubenswrapper[4921]: I0103 03:58:08.232840 4921 generic.go:334] "Generic (PLEG): container finished" podID="08a322d7-503b-498f-99cb-149fd7032129" containerID="1a5a4f91a65be9cdd0f1eb10f98afb923634e2d0ef10ed7fd30363e01735e1f8" exitCode=0 Jan 03 03:58:08 crc kubenswrapper[4921]: I0103 03:58:08.232884 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerDied","Data":"1a5a4f91a65be9cdd0f1eb10f98afb923634e2d0ef10ed7fd30363e01735e1f8"} Jan 03 03:58:11 crc kubenswrapper[4921]: I0103 03:58:11.282836 4921 generic.go:334] "Generic (PLEG): container finished" podID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerID="cbc8db5ecf5e62cd68d2dc5927977a9308b4bc4304d0390e270801631d4f1dce" exitCode=0 Jan 03 03:58:11 crc kubenswrapper[4921]: I0103 03:58:11.282947 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" event={"ID":"4ca0e16f-8e0c-455a-9767-951056bd29cb","Type":"ContainerDied","Data":"cbc8db5ecf5e62cd68d2dc5927977a9308b4bc4304d0390e270801631d4f1dce"} Jan 03 03:58:11 crc kubenswrapper[4921]: I0103 03:58:11.286773 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerStarted","Data":"4ae1cc421c22eda34b6d53468d5105d0325901a033f6b5f928d0c1572fea3dc7"} Jan 03 03:58:12 crc kubenswrapper[4921]: I0103 03:58:12.296233 4921 generic.go:334] "Generic (PLEG): container finished" podID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerID="7dd9c898deafb637ee9c16162ffb3b89cbd1c7890daede04a7bc72c5335d438d" exitCode=0 Jan 03 03:58:12 crc kubenswrapper[4921]: I0103 03:58:12.296729 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" event={"ID":"4ca0e16f-8e0c-455a-9767-951056bd29cb","Type":"ContainerDied","Data":"7dd9c898deafb637ee9c16162ffb3b89cbd1c7890daede04a7bc72c5335d438d"} Jan 03 03:58:12 crc kubenswrapper[4921]: I0103 03:58:12.322862 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xzbrh" podStartSLOduration=4.181426671 podStartE2EDuration="8.3228151s" podCreationTimestamp="2026-01-03 03:58:04 +0000 UTC" firstStartedPulling="2026-01-03 03:58:06.215344844 +0000 UTC m=+1021.826771668" lastFinishedPulling="2026-01-03 03:58:10.356733263 +0000 UTC m=+1025.968160097" observedRunningTime="2026-01-03 03:58:11.339344904 +0000 UTC m=+1026.950771728" watchObservedRunningTime="2026-01-03 03:58:12.3228151 +0000 UTC m=+1027.934241964" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.556005 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.663851 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h44h\" (UniqueName: \"kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h\") pod \"4ca0e16f-8e0c-455a-9767-951056bd29cb\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.664031 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle\") pod \"4ca0e16f-8e0c-455a-9767-951056bd29cb\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.664091 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util\") pod \"4ca0e16f-8e0c-455a-9767-951056bd29cb\" (UID: \"4ca0e16f-8e0c-455a-9767-951056bd29cb\") " Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.664850 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle" (OuterVolumeSpecName: "bundle") pod "4ca0e16f-8e0c-455a-9767-951056bd29cb" (UID: "4ca0e16f-8e0c-455a-9767-951056bd29cb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.669660 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h" (OuterVolumeSpecName: "kube-api-access-9h44h") pod "4ca0e16f-8e0c-455a-9767-951056bd29cb" (UID: "4ca0e16f-8e0c-455a-9767-951056bd29cb"). InnerVolumeSpecName "kube-api-access-9h44h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.674778 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util" (OuterVolumeSpecName: "util") pod "4ca0e16f-8e0c-455a-9767-951056bd29cb" (UID: "4ca0e16f-8e0c-455a-9767-951056bd29cb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.765186 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.765230 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h44h\" (UniqueName: \"kubernetes.io/projected/4ca0e16f-8e0c-455a-9767-951056bd29cb-kube-api-access-9h44h\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:13 crc kubenswrapper[4921]: I0103 03:58:13.765245 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ca0e16f-8e0c-455a-9767-951056bd29cb-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:14 crc kubenswrapper[4921]: I0103 03:58:14.313088 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" event={"ID":"4ca0e16f-8e0c-455a-9767-951056bd29cb","Type":"ContainerDied","Data":"f03340c269b4a0b18c48db63b1dc35624e4766c8cd4cf7faa21b1ecc925633a8"} Jan 03 03:58:14 crc kubenswrapper[4921]: I0103 03:58:14.313138 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f03340c269b4a0b18c48db63b1dc35624e4766c8cd4cf7faa21b1ecc925633a8" Jan 03 03:58:14 crc kubenswrapper[4921]: I0103 03:58:14.313220 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k" Jan 03 03:58:15 crc kubenswrapper[4921]: I0103 03:58:15.319211 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:15 crc kubenswrapper[4921]: I0103 03:58:15.319579 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:16 crc kubenswrapper[4921]: I0103 03:58:16.387326 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xzbrh" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="registry-server" probeResult="failure" output=< Jan 03 03:58:16 crc kubenswrapper[4921]: timeout: failed to connect service ":50051" within 1s Jan 03 03:58:16 crc kubenswrapper[4921]: > Jan 03 03:58:18 crc kubenswrapper[4921]: I0103 03:58:18.346558 4921 generic.go:334] "Generic (PLEG): container finished" podID="e3b45f46-86ca-4ffb-b451-00d8631b9aaa" containerID="bc1f1683c7e0e98827171a6cb4b13587562267bfccbe4b7c822e7adc5de9b78c" exitCode=0 Jan 03 03:58:18 crc kubenswrapper[4921]: I0103 03:58:18.346763 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"e3b45f46-86ca-4ffb-b451-00d8631b9aaa","Type":"ContainerDied","Data":"bc1f1683c7e0e98827171a6cb4b13587562267bfccbe4b7c822e7adc5de9b78c"} Jan 03 03:58:19 crc kubenswrapper[4921]: I0103 03:58:19.356193 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"e3b45f46-86ca-4ffb-b451-00d8631b9aaa","Type":"ContainerStarted","Data":"01bbd5581c47c3158a48a19f71e6819a67a84380c943f77617202d49c72f296c"} Jan 03 03:58:19 crc kubenswrapper[4921]: I0103 03:58:19.356719 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:58:19 crc kubenswrapper[4921]: I0103 03:58:19.392599 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.777422046 podStartE2EDuration="53.392569969s" podCreationTimestamp="2026-01-03 03:57:26 +0000 UTC" firstStartedPulling="2026-01-03 03:57:27.73251767 +0000 UTC m=+983.343944504" lastFinishedPulling="2026-01-03 03:57:43.347665603 +0000 UTC m=+998.959092427" observedRunningTime="2026-01-03 03:58:19.382875103 +0000 UTC m=+1034.994301917" watchObservedRunningTime="2026-01-03 03:58:19.392569969 +0000 UTC m=+1035.003996833" Jan 03 03:58:25 crc kubenswrapper[4921]: I0103 03:58:25.385238 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:25 crc kubenswrapper[4921]: I0103 03:58:25.440253 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:28 crc kubenswrapper[4921]: I0103 03:58:28.782127 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:28 crc kubenswrapper[4921]: I0103 03:58:28.783019 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xzbrh" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="registry-server" containerID="cri-o://4ae1cc421c22eda34b6d53468d5105d0325901a033f6b5f928d0c1572fea3dc7" gracePeriod=2 Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.422613 4921 generic.go:334] "Generic (PLEG): container finished" podID="08a322d7-503b-498f-99cb-149fd7032129" containerID="4ae1cc421c22eda34b6d53468d5105d0325901a033f6b5f928d0c1572fea3dc7" exitCode=0 Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.422658 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerDied","Data":"4ae1cc421c22eda34b6d53468d5105d0325901a033f6b5f928d0c1572fea3dc7"} Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.669161 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.794955 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content\") pod \"08a322d7-503b-498f-99cb-149fd7032129\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.795042 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp6zh\" (UniqueName: \"kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh\") pod \"08a322d7-503b-498f-99cb-149fd7032129\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.795093 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities\") pod \"08a322d7-503b-498f-99cb-149fd7032129\" (UID: \"08a322d7-503b-498f-99cb-149fd7032129\") " Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.796552 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities" (OuterVolumeSpecName: "utilities") pod "08a322d7-503b-498f-99cb-149fd7032129" (UID: "08a322d7-503b-498f-99cb-149fd7032129"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.809595 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh" (OuterVolumeSpecName: "kube-api-access-qp6zh") pod "08a322d7-503b-498f-99cb-149fd7032129" (UID: "08a322d7-503b-498f-99cb-149fd7032129"). InnerVolumeSpecName "kube-api-access-qp6zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.897474 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp6zh\" (UniqueName: \"kubernetes.io/projected/08a322d7-503b-498f-99cb-149fd7032129-kube-api-access-qp6zh\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.897537 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.920817 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08a322d7-503b-498f-99cb-149fd7032129" (UID: "08a322d7-503b-498f-99cb-149fd7032129"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:58:29 crc kubenswrapper[4921]: I0103 03:58:29.999600 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a322d7-503b-498f-99cb-149fd7032129-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.431897 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xzbrh" event={"ID":"08a322d7-503b-498f-99cb-149fd7032129","Type":"ContainerDied","Data":"5306f88ddf819c345c2945ab8d14df792cf5a0c7113dd436b42742716d95343b"} Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.431947 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xzbrh" Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.431969 4921 scope.go:117] "RemoveContainer" containerID="4ae1cc421c22eda34b6d53468d5105d0325901a033f6b5f928d0c1572fea3dc7" Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.460087 4921 scope.go:117] "RemoveContainer" containerID="1a5a4f91a65be9cdd0f1eb10f98afb923634e2d0ef10ed7fd30363e01735e1f8" Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.466397 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.477774 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xzbrh"] Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.514548 4921 scope.go:117] "RemoveContainer" containerID="714c78e77d9f76fb74a16602b38cde0b0d9788209f65ceac702214f2bc2e3507" Jan 03 03:58:30 crc kubenswrapper[4921]: I0103 03:58:30.904513 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08a322d7-503b-498f-99cb-149fd7032129" path="/var/lib/kubelet/pods/08a322d7-503b-498f-99cb-149fd7032129/volumes" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.527886 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb"] Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528390 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="pull" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528402 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="pull" Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528411 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="extract-content" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528418 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="extract-content" Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528438 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="util" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528444 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="util" Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528456 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="extract-utilities" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528464 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="extract-utilities" Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528472 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="registry-server" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528477 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="registry-server" Jan 03 03:58:34 crc kubenswrapper[4921]: E0103 03:58:34.528484 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="extract" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528490 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="extract" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528609 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a322d7-503b-498f-99cb-149fd7032129" containerName="registry-server" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.528625 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca0e16f-8e0c-455a-9767-951056bd29cb" containerName="extract" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.529286 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.531595 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rzpld" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.531597 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.542985 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb"] Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.573578 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-apiservice-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.573656 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-webhook-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.573677 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrrmd\" (UniqueName: \"kubernetes.io/projected/51565a9f-65ee-4007-b46d-bf3100a220a6-kube-api-access-mrrmd\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.674798 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-apiservice-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.674957 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-webhook-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.674998 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrrmd\" (UniqueName: \"kubernetes.io/projected/51565a9f-65ee-4007-b46d-bf3100a220a6-kube-api-access-mrrmd\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.690884 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-apiservice-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.691175 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/51565a9f-65ee-4007-b46d-bf3100a220a6-webhook-cert\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.702892 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrrmd\" (UniqueName: \"kubernetes.io/projected/51565a9f-65ee-4007-b46d-bf3100a220a6-kube-api-access-mrrmd\") pod \"keystone-operator-controller-manager-6486d9c577-kpxkb\" (UID: \"51565a9f-65ee-4007-b46d-bf3100a220a6\") " pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:34 crc kubenswrapper[4921]: I0103 03:58:34.861253 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:35 crc kubenswrapper[4921]: I0103 03:58:35.148522 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb"] Jan 03 03:58:35 crc kubenswrapper[4921]: W0103 03:58:35.153415 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51565a9f_65ee_4007_b46d_bf3100a220a6.slice/crio-79138d9fb7d4943b027fed60baed1e8df049f9581491f7895bbf1fdd3b30fb3b WatchSource:0}: Error finding container 79138d9fb7d4943b027fed60baed1e8df049f9581491f7895bbf1fdd3b30fb3b: Status 404 returned error can't find the container with id 79138d9fb7d4943b027fed60baed1e8df049f9581491f7895bbf1fdd3b30fb3b Jan 03 03:58:35 crc kubenswrapper[4921]: I0103 03:58:35.156347 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 03:58:35 crc kubenswrapper[4921]: I0103 03:58:35.472427 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" event={"ID":"51565a9f-65ee-4007-b46d-bf3100a220a6","Type":"ContainerStarted","Data":"79138d9fb7d4943b027fed60baed1e8df049f9581491f7895bbf1fdd3b30fb3b"} Jan 03 03:58:37 crc kubenswrapper[4921]: I0103 03:58:37.485419 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Jan 03 03:58:37 crc kubenswrapper[4921]: I0103 03:58:37.487984 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" event={"ID":"51565a9f-65ee-4007-b46d-bf3100a220a6","Type":"ContainerStarted","Data":"8546258f5b867a95d217329cb5af75292a160d44b1abf8f2ec909640ee9700bc"} Jan 03 03:58:37 crc kubenswrapper[4921]: I0103 03:58:37.488040 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" event={"ID":"51565a9f-65ee-4007-b46d-bf3100a220a6","Type":"ContainerStarted","Data":"9b0ac50b3ddb54abe2f9afed1026fdf6d540893f0a31b0d71fe621b21b9323ea"} Jan 03 03:58:37 crc kubenswrapper[4921]: I0103 03:58:37.488219 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:37 crc kubenswrapper[4921]: I0103 03:58:37.596575 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" podStartSLOduration=1.5848604910000001 podStartE2EDuration="3.596558495s" podCreationTimestamp="2026-01-03 03:58:34 +0000 UTC" firstStartedPulling="2026-01-03 03:58:35.155953661 +0000 UTC m=+1050.767380505" lastFinishedPulling="2026-01-03 03:58:37.167651675 +0000 UTC m=+1052.779078509" observedRunningTime="2026-01-03 03:58:37.593903323 +0000 UTC m=+1053.205330177" watchObservedRunningTime="2026-01-03 03:58:37.596558495 +0000 UTC m=+1053.207985319" Jan 03 03:58:44 crc kubenswrapper[4921]: I0103 03:58:44.867182 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-6486d9c577-kpxkb" Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.176301 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-bj9mx"] Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.178730 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.183701 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-bj9mx"] Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.375508 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsw27\" (UniqueName: \"kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27\") pod \"keystone-db-create-bj9mx\" (UID: \"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261\") " pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.476614 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsw27\" (UniqueName: \"kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27\") pod \"keystone-db-create-bj9mx\" (UID: \"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261\") " pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.511819 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsw27\" (UniqueName: \"kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27\") pod \"keystone-db-create-bj9mx\" (UID: \"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261\") " pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:47 crc kubenswrapper[4921]: I0103 03:58:47.797805 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:48 crc kubenswrapper[4921]: I0103 03:58:48.229432 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-bj9mx"] Jan 03 03:58:48 crc kubenswrapper[4921]: I0103 03:58:48.576084 4921 generic.go:334] "Generic (PLEG): container finished" podID="ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" containerID="5a0c7beba4b4884779fa1b1f4ae3bc0a21477bcf48d10362755b7f3e6df03e7f" exitCode=0 Jan 03 03:58:48 crc kubenswrapper[4921]: I0103 03:58:48.576173 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-bj9mx" event={"ID":"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261","Type":"ContainerDied","Data":"5a0c7beba4b4884779fa1b1f4ae3bc0a21477bcf48d10362755b7f3e6df03e7f"} Jan 03 03:58:48 crc kubenswrapper[4921]: I0103 03:58:48.576545 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-bj9mx" event={"ID":"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261","Type":"ContainerStarted","Data":"438eecf4baf21d19622779dde2d44bf5ca000dc93706d18d0bf4a56c57fc742c"} Jan 03 03:58:49 crc kubenswrapper[4921]: I0103 03:58:49.870711 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.016517 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsw27\" (UniqueName: \"kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27\") pod \"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261\" (UID: \"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261\") " Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.022224 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27" (OuterVolumeSpecName: "kube-api-access-fsw27") pod "ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" (UID: "ee9d2b0d-28d2-4d99-a67c-71af6d2ee261"). InnerVolumeSpecName "kube-api-access-fsw27". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.118074 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsw27\" (UniqueName: \"kubernetes.io/projected/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261-kube-api-access-fsw27\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.190020 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:50 crc kubenswrapper[4921]: E0103 03:58:50.190315 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" containerName="mariadb-database-create" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.190333 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" containerName="mariadb-database-create" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.190449 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" containerName="mariadb-database-create" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.190998 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.193044 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-c2s8k" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.197813 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.320494 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swb2s\" (UniqueName: \"kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s\") pod \"horizon-operator-index-wvpnh\" (UID: \"f0a1b8c9-5b29-49e6-9f68-afe8357bca11\") " pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.422482 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swb2s\" (UniqueName: \"kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s\") pod \"horizon-operator-index-wvpnh\" (UID: \"f0a1b8c9-5b29-49e6-9f68-afe8357bca11\") " pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.458493 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swb2s\" (UniqueName: \"kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s\") pod \"horizon-operator-index-wvpnh\" (UID: \"f0a1b8c9-5b29-49e6-9f68-afe8357bca11\") " pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.515743 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.599040 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-bj9mx" event={"ID":"ee9d2b0d-28d2-4d99-a67c-71af6d2ee261","Type":"ContainerDied","Data":"438eecf4baf21d19622779dde2d44bf5ca000dc93706d18d0bf4a56c57fc742c"} Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.599444 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="438eecf4baf21d19622779dde2d44bf5ca000dc93706d18d0bf4a56c57fc742c" Jan 03 03:58:50 crc kubenswrapper[4921]: I0103 03:58:50.599121 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-bj9mx" Jan 03 03:58:51 crc kubenswrapper[4921]: I0103 03:58:51.028895 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:51 crc kubenswrapper[4921]: W0103 03:58:51.036931 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0a1b8c9_5b29_49e6_9f68_afe8357bca11.slice/crio-0921c063594a2ccc186d59d52f19fd3adbfe616ef5be21d52a532ee33e740a52 WatchSource:0}: Error finding container 0921c063594a2ccc186d59d52f19fd3adbfe616ef5be21d52a532ee33e740a52: Status 404 returned error can't find the container with id 0921c063594a2ccc186d59d52f19fd3adbfe616ef5be21d52a532ee33e740a52 Jan 03 03:58:51 crc kubenswrapper[4921]: I0103 03:58:51.608924 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-wvpnh" event={"ID":"f0a1b8c9-5b29-49e6-9f68-afe8357bca11","Type":"ContainerStarted","Data":"0921c063594a2ccc186d59d52f19fd3adbfe616ef5be21d52a532ee33e740a52"} Jan 03 03:58:52 crc kubenswrapper[4921]: I0103 03:58:52.616240 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-wvpnh" event={"ID":"f0a1b8c9-5b29-49e6-9f68-afe8357bca11","Type":"ContainerStarted","Data":"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c"} Jan 03 03:58:52 crc kubenswrapper[4921]: I0103 03:58:52.649597 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-wvpnh" podStartSLOduration=1.5699932749999999 podStartE2EDuration="2.649571501s" podCreationTimestamp="2026-01-03 03:58:50 +0000 UTC" firstStartedPulling="2026-01-03 03:58:51.041311425 +0000 UTC m=+1066.652738249" lastFinishedPulling="2026-01-03 03:58:52.120889611 +0000 UTC m=+1067.732316475" observedRunningTime="2026-01-03 03:58:52.636160347 +0000 UTC m=+1068.247587171" watchObservedRunningTime="2026-01-03 03:58:52.649571501 +0000 UTC m=+1068.260998345" Jan 03 03:58:52 crc kubenswrapper[4921]: I0103 03:58:52.994567 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-54hh2"] Jan 03 03:58:52 crc kubenswrapper[4921]: I0103 03:58:52.995937 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.000000 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-kzg5m" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.007093 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-54hh2"] Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.070179 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f949k\" (UniqueName: \"kubernetes.io/projected/d335cc0f-f8c6-4810-9113-cb8b934d142e-kube-api-access-f949k\") pod \"swift-operator-index-54hh2\" (UID: \"d335cc0f-f8c6-4810-9113-cb8b934d142e\") " pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.171758 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f949k\" (UniqueName: \"kubernetes.io/projected/d335cc0f-f8c6-4810-9113-cb8b934d142e-kube-api-access-f949k\") pod \"swift-operator-index-54hh2\" (UID: \"d335cc0f-f8c6-4810-9113-cb8b934d142e\") " pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.205805 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f949k\" (UniqueName: \"kubernetes.io/projected/d335cc0f-f8c6-4810-9113-cb8b934d142e-kube-api-access-f949k\") pod \"swift-operator-index-54hh2\" (UID: \"d335cc0f-f8c6-4810-9113-cb8b934d142e\") " pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.321814 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.568693 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-54hh2"] Jan 03 03:58:53 crc kubenswrapper[4921]: W0103 03:58:53.576945 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd335cc0f_f8c6_4810_9113_cb8b934d142e.slice/crio-1a193d1fdb1d97fa1735d2e6fbc8f318e883779b7668ae131745c1aa977c3d70 WatchSource:0}: Error finding container 1a193d1fdb1d97fa1735d2e6fbc8f318e883779b7668ae131745c1aa977c3d70: Status 404 returned error can't find the container with id 1a193d1fdb1d97fa1735d2e6fbc8f318e883779b7668ae131745c1aa977c3d70 Jan 03 03:58:53 crc kubenswrapper[4921]: I0103 03:58:53.622429 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-54hh2" event={"ID":"d335cc0f-f8c6-4810-9113-cb8b934d142e","Type":"ContainerStarted","Data":"1a193d1fdb1d97fa1735d2e6fbc8f318e883779b7668ae131745c1aa977c3d70"} Jan 03 03:58:55 crc kubenswrapper[4921]: I0103 03:58:55.645931 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-54hh2" event={"ID":"d335cc0f-f8c6-4810-9113-cb8b934d142e","Type":"ContainerStarted","Data":"e97f493aa6ddcc00ab1bec2030891ad7a91c06c574e5838ccdd5ca63e1025d5a"} Jan 03 03:58:55 crc kubenswrapper[4921]: I0103 03:58:55.666530 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-54hh2" podStartSLOduration=2.655044656 podStartE2EDuration="3.66650376s" podCreationTimestamp="2026-01-03 03:58:52 +0000 UTC" firstStartedPulling="2026-01-03 03:58:53.57910007 +0000 UTC m=+1069.190526914" lastFinishedPulling="2026-01-03 03:58:54.590559174 +0000 UTC m=+1070.201986018" observedRunningTime="2026-01-03 03:58:55.662873562 +0000 UTC m=+1071.274300406" watchObservedRunningTime="2026-01-03 03:58:55.66650376 +0000 UTC m=+1071.277930594" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.179968 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.180693 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/horizon-operator-index-wvpnh" podUID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" containerName="registry-server" containerID="cri-o://a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c" gracePeriod=2 Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.642033 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.654051 4921 generic.go:334] "Generic (PLEG): container finished" podID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" containerID="a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c" exitCode=0 Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.654125 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-wvpnh" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.654187 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-wvpnh" event={"ID":"f0a1b8c9-5b29-49e6-9f68-afe8357bca11","Type":"ContainerDied","Data":"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c"} Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.654247 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-wvpnh" event={"ID":"f0a1b8c9-5b29-49e6-9f68-afe8357bca11","Type":"ContainerDied","Data":"0921c063594a2ccc186d59d52f19fd3adbfe616ef5be21d52a532ee33e740a52"} Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.654292 4921 scope.go:117] "RemoveContainer" containerID="a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.690515 4921 scope.go:117] "RemoveContainer" containerID="a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c" Jan 03 03:58:56 crc kubenswrapper[4921]: E0103 03:58:56.691830 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c\": container with ID starting with a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c not found: ID does not exist" containerID="a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.691958 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c"} err="failed to get container status \"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c\": rpc error: code = NotFound desc = could not find container \"a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c\": container with ID starting with a57f8d8f2539d7023384584d3c72d5ccca02579c5f19f5dbf08e26b44510b87c not found: ID does not exist" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.750751 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swb2s\" (UniqueName: \"kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s\") pod \"f0a1b8c9-5b29-49e6-9f68-afe8357bca11\" (UID: \"f0a1b8c9-5b29-49e6-9f68-afe8357bca11\") " Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.759511 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s" (OuterVolumeSpecName: "kube-api-access-swb2s") pod "f0a1b8c9-5b29-49e6-9f68-afe8357bca11" (UID: "f0a1b8c9-5b29-49e6-9f68-afe8357bca11"). InnerVolumeSpecName "kube-api-access-swb2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.852803 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swb2s\" (UniqueName: \"kubernetes.io/projected/f0a1b8c9-5b29-49e6-9f68-afe8357bca11-kube-api-access-swb2s\") on node \"crc\" DevicePath \"\"" Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.991012 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:56 crc kubenswrapper[4921]: I0103 03:58:56.998244 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/horizon-operator-index-wvpnh"] Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.007238 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-rjwtj"] Jan 03 03:58:57 crc kubenswrapper[4921]: E0103 03:58:57.007576 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" containerName="registry-server" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.007591 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" containerName="registry-server" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.007715 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" containerName="registry-server" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.008111 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.010436 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-c2s8k" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.013231 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-rjwtj"] Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.055331 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhzjt\" (UniqueName: \"kubernetes.io/projected/69216a14-e1e4-4b97-8faf-d858a829143d-kube-api-access-zhzjt\") pod \"horizon-operator-index-rjwtj\" (UID: \"69216a14-e1e4-4b97-8faf-d858a829143d\") " pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.073530 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-9b09-account-create-5nc6d"] Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.074819 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.079682 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.099502 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-9b09-account-create-5nc6d"] Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.156199 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhzjt\" (UniqueName: \"kubernetes.io/projected/69216a14-e1e4-4b97-8faf-d858a829143d-kube-api-access-zhzjt\") pod \"horizon-operator-index-rjwtj\" (UID: \"69216a14-e1e4-4b97-8faf-d858a829143d\") " pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.174222 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhzjt\" (UniqueName: \"kubernetes.io/projected/69216a14-e1e4-4b97-8faf-d858a829143d-kube-api-access-zhzjt\") pod \"horizon-operator-index-rjwtj\" (UID: \"69216a14-e1e4-4b97-8faf-d858a829143d\") " pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.257355 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdspg\" (UniqueName: \"kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg\") pod \"keystone-9b09-account-create-5nc6d\" (UID: \"1ad2843b-da3a-4404-af04-6b552cddfa6d\") " pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.322247 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.358685 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdspg\" (UniqueName: \"kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg\") pod \"keystone-9b09-account-create-5nc6d\" (UID: \"1ad2843b-da3a-4404-af04-6b552cddfa6d\") " pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.378786 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdspg\" (UniqueName: \"kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg\") pod \"keystone-9b09-account-create-5nc6d\" (UID: \"1ad2843b-da3a-4404-af04-6b552cddfa6d\") " pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.392562 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.881221 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-rjwtj"] Jan 03 03:58:57 crc kubenswrapper[4921]: I0103 03:58:57.905053 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-9b09-account-create-5nc6d"] Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.669743 4921 generic.go:334] "Generic (PLEG): container finished" podID="1ad2843b-da3a-4404-af04-6b552cddfa6d" containerID="e27ec8fc352ae5fa4499b9400a934e3e933091d3190d62428162ab14e0ddbc36" exitCode=0 Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.669874 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" event={"ID":"1ad2843b-da3a-4404-af04-6b552cddfa6d","Type":"ContainerDied","Data":"e27ec8fc352ae5fa4499b9400a934e3e933091d3190d62428162ab14e0ddbc36"} Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.670131 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" event={"ID":"1ad2843b-da3a-4404-af04-6b552cddfa6d","Type":"ContainerStarted","Data":"5b1c6243ad6bf6e77550eef530bb112e18f203da9d6b2e617e5adf3da10f60df"} Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.671797 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-rjwtj" event={"ID":"69216a14-e1e4-4b97-8faf-d858a829143d","Type":"ContainerStarted","Data":"30fa4628eca4a0da8f4260db61a8ef71db06c3a6ddc24803e428d926215d403a"} Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.671831 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-rjwtj" event={"ID":"69216a14-e1e4-4b97-8faf-d858a829143d","Type":"ContainerStarted","Data":"fb4eaad62da9d2477bc67000e93c9cc10070e157df524f764c2171b1a545b252"} Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.700908 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-rjwtj" podStartSLOduration=2.19859688 podStartE2EDuration="2.700886414s" podCreationTimestamp="2026-01-03 03:58:56 +0000 UTC" firstStartedPulling="2026-01-03 03:58:57.890579343 +0000 UTC m=+1073.502006167" lastFinishedPulling="2026-01-03 03:58:58.392868867 +0000 UTC m=+1074.004295701" observedRunningTime="2026-01-03 03:58:58.69815393 +0000 UTC m=+1074.309580754" watchObservedRunningTime="2026-01-03 03:58:58.700886414 +0000 UTC m=+1074.312313238" Jan 03 03:58:58 crc kubenswrapper[4921]: I0103 03:58:58.894192 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a1b8c9-5b29-49e6-9f68-afe8357bca11" path="/var/lib/kubelet/pods/f0a1b8c9-5b29-49e6-9f68-afe8357bca11/volumes" Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.004747 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.196104 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdspg\" (UniqueName: \"kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg\") pod \"1ad2843b-da3a-4404-af04-6b552cddfa6d\" (UID: \"1ad2843b-da3a-4404-af04-6b552cddfa6d\") " Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.201309 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg" (OuterVolumeSpecName: "kube-api-access-zdspg") pod "1ad2843b-da3a-4404-af04-6b552cddfa6d" (UID: "1ad2843b-da3a-4404-af04-6b552cddfa6d"). InnerVolumeSpecName "kube-api-access-zdspg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.298116 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdspg\" (UniqueName: \"kubernetes.io/projected/1ad2843b-da3a-4404-af04-6b552cddfa6d-kube-api-access-zdspg\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.689553 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" event={"ID":"1ad2843b-da3a-4404-af04-6b552cddfa6d","Type":"ContainerDied","Data":"5b1c6243ad6bf6e77550eef530bb112e18f203da9d6b2e617e5adf3da10f60df"} Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.689588 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b1c6243ad6bf6e77550eef530bb112e18f203da9d6b2e617e5adf3da10f60df" Jan 03 03:59:00 crc kubenswrapper[4921]: I0103 03:59:00.689696 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-9b09-account-create-5nc6d" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.660011 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-45jwt"] Jan 03 03:59:02 crc kubenswrapper[4921]: E0103 03:59:02.660417 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ad2843b-da3a-4404-af04-6b552cddfa6d" containerName="mariadb-account-create" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.660435 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ad2843b-da3a-4404-af04-6b552cddfa6d" containerName="mariadb-account-create" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.660622 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ad2843b-da3a-4404-af04-6b552cddfa6d" containerName="mariadb-account-create" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.661159 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.662990 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.663945 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.664394 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-7c2nh" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.664745 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.674468 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-45jwt"] Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.836109 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.836230 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdvqj\" (UniqueName: \"kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.938031 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.938741 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdvqj\" (UniqueName: \"kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.942307 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.963222 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdvqj\" (UniqueName: \"kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj\") pod \"keystone-db-sync-45jwt\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:02 crc kubenswrapper[4921]: I0103 03:59:02.978149 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.322667 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.322929 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.368111 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.432755 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-45jwt"] Jan 03 03:59:03 crc kubenswrapper[4921]: W0103 03:59:03.443006 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb022c441_4cd8_4f96_a93b_cd194a4a505b.slice/crio-0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9 WatchSource:0}: Error finding container 0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9: Status 404 returned error can't find the container with id 0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9 Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.719707 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-45jwt" event={"ID":"b022c441-4cd8-4f96-a93b-cd194a4a505b","Type":"ContainerStarted","Data":"0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9"} Jan 03 03:59:03 crc kubenswrapper[4921]: I0103 03:59:03.768644 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-54hh2" Jan 03 03:59:07 crc kubenswrapper[4921]: I0103 03:59:07.323483 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:59:07 crc kubenswrapper[4921]: I0103 03:59:07.324108 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:59:07 crc kubenswrapper[4921]: I0103 03:59:07.353938 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:59:07 crc kubenswrapper[4921]: I0103 03:59:07.773002 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-rjwtj" Jan 03 03:59:15 crc kubenswrapper[4921]: I0103 03:59:15.809970 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-45jwt" event={"ID":"b022c441-4cd8-4f96-a93b-cd194a4a505b","Type":"ContainerStarted","Data":"9f1294aa84cde1c458fb6eeafea7bf173f557e036e7324c31fa60be3884caf6b"} Jan 03 03:59:15 crc kubenswrapper[4921]: I0103 03:59:15.834804 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-45jwt" podStartSLOduration=2.649080496 podStartE2EDuration="13.834774724s" podCreationTimestamp="2026-01-03 03:59:02 +0000 UTC" firstStartedPulling="2026-01-03 03:59:03.447202319 +0000 UTC m=+1079.058629153" lastFinishedPulling="2026-01-03 03:59:14.632896557 +0000 UTC m=+1090.244323381" observedRunningTime="2026-01-03 03:59:15.832521912 +0000 UTC m=+1091.443948756" watchObservedRunningTime="2026-01-03 03:59:15.834774724 +0000 UTC m=+1091.446201548" Jan 03 03:59:18 crc kubenswrapper[4921]: I0103 03:59:18.833863 4921 generic.go:334] "Generic (PLEG): container finished" podID="b022c441-4cd8-4f96-a93b-cd194a4a505b" containerID="9f1294aa84cde1c458fb6eeafea7bf173f557e036e7324c31fa60be3884caf6b" exitCode=0 Jan 03 03:59:18 crc kubenswrapper[4921]: I0103 03:59:18.833979 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-45jwt" event={"ID":"b022c441-4cd8-4f96-a93b-cd194a4a505b","Type":"ContainerDied","Data":"9f1294aa84cde1c458fb6eeafea7bf173f557e036e7324c31fa60be3884caf6b"} Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.237194 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.406462 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdvqj\" (UniqueName: \"kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj\") pod \"b022c441-4cd8-4f96-a93b-cd194a4a505b\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.406599 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data\") pod \"b022c441-4cd8-4f96-a93b-cd194a4a505b\" (UID: \"b022c441-4cd8-4f96-a93b-cd194a4a505b\") " Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.414632 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj" (OuterVolumeSpecName: "kube-api-access-bdvqj") pod "b022c441-4cd8-4f96-a93b-cd194a4a505b" (UID: "b022c441-4cd8-4f96-a93b-cd194a4a505b"). InnerVolumeSpecName "kube-api-access-bdvqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.434140 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data" (OuterVolumeSpecName: "config-data") pod "b022c441-4cd8-4f96-a93b-cd194a4a505b" (UID: "b022c441-4cd8-4f96-a93b-cd194a4a505b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.508219 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdvqj\" (UniqueName: \"kubernetes.io/projected/b022c441-4cd8-4f96-a93b-cd194a4a505b-kube-api-access-bdvqj\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.508257 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b022c441-4cd8-4f96-a93b-cd194a4a505b-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.852788 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-45jwt" event={"ID":"b022c441-4cd8-4f96-a93b-cd194a4a505b","Type":"ContainerDied","Data":"0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9"} Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.853115 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a0f1f2a05febc7eabc5321a0f1c17c9f109a01d191b4affdccf4c68a26de4b9" Jan 03 03:59:20 crc kubenswrapper[4921]: I0103 03:59:20.852896 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-45jwt" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.083986 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6vmcw"] Jan 03 03:59:21 crc kubenswrapper[4921]: E0103 03:59:21.084368 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b022c441-4cd8-4f96-a93b-cd194a4a505b" containerName="keystone-db-sync" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.084395 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b022c441-4cd8-4f96-a93b-cd194a4a505b" containerName="keystone-db-sync" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.084630 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="b022c441-4cd8-4f96-a93b-cd194a4a505b" containerName="keystone-db-sync" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.085337 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.088013 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-7c2nh" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.088180 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.088337 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.091353 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.100116 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6vmcw"] Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.219471 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.219579 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.219607 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.219688 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.219743 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.321591 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.321684 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.321749 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.321783 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.321820 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.325928 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.326059 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.327471 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.337164 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.351109 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd\") pod \"keystone-bootstrap-6vmcw\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.407822 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.827756 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6vmcw"] Jan 03 03:59:21 crc kubenswrapper[4921]: I0103 03:59:21.866521 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" event={"ID":"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80","Type":"ContainerStarted","Data":"970d0d7c8ab360e1d2c6cc50040c3e710bf6f4cfd39e48b0e01ca5e7b2c2bf0b"} Jan 03 03:59:22 crc kubenswrapper[4921]: I0103 03:59:22.880144 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" event={"ID":"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80","Type":"ContainerStarted","Data":"6d6c144736539e6b6cea09efc8064df216c26a5e5837bc194e0ed3984ea93603"} Jan 03 03:59:22 crc kubenswrapper[4921]: I0103 03:59:22.913647 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" podStartSLOduration=1.9136224670000002 podStartE2EDuration="1.913622467s" podCreationTimestamp="2026-01-03 03:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:59:22.902876145 +0000 UTC m=+1098.514303009" watchObservedRunningTime="2026-01-03 03:59:22.913622467 +0000 UTC m=+1098.525049321" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.439875 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h"] Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.441567 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.443957 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.453132 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h"] Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.572941 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.573025 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.573095 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79wsp\" (UniqueName: \"kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.674912 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.675000 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.675069 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79wsp\" (UniqueName: \"kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.675712 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.675712 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.708324 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79wsp\" (UniqueName: \"kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp\") pod \"f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:23 crc kubenswrapper[4921]: I0103 03:59:23.762791 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.047878 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h"] Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.427175 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q"] Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.428664 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.448220 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q"] Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.586185 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.586235 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.586264 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmnsb\" (UniqueName: \"kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.687875 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.687926 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.687956 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmnsb\" (UniqueName: \"kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.688772 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.688866 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.710962 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmnsb\" (UniqueName: \"kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb\") pod \"5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.744856 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.905865 4921 generic.go:334] "Generic (PLEG): container finished" podID="34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" containerID="6d6c144736539e6b6cea09efc8064df216c26a5e5837bc194e0ed3984ea93603" exitCode=0 Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.905960 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" event={"ID":"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80","Type":"ContainerDied","Data":"6d6c144736539e6b6cea09efc8064df216c26a5e5837bc194e0ed3984ea93603"} Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.913258 4921 generic.go:334] "Generic (PLEG): container finished" podID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerID="746aa965860b337502f1a8ec74b84044aec2df8173d1d5331ca462b239b8923e" exitCode=0 Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.913316 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" event={"ID":"876c356f-d130-4a6f-b888-5a32ab0d83da","Type":"ContainerDied","Data":"746aa965860b337502f1a8ec74b84044aec2df8173d1d5331ca462b239b8923e"} Jan 03 03:59:24 crc kubenswrapper[4921]: I0103 03:59:24.913339 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" event={"ID":"876c356f-d130-4a6f-b888-5a32ab0d83da","Type":"ContainerStarted","Data":"7ee2c8f8663c23f40b87cd8832d5e1efa2d35034c65399dea1e223dc9df327ac"} Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.181787 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q"] Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.919077 4921 generic.go:334] "Generic (PLEG): container finished" podID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerID="1fa6235a02aea39960f06f00ebc06bee93ada17587e041ac90a1c057acbe505f" exitCode=0 Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.919177 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" event={"ID":"876c356f-d130-4a6f-b888-5a32ab0d83da","Type":"ContainerDied","Data":"1fa6235a02aea39960f06f00ebc06bee93ada17587e041ac90a1c057acbe505f"} Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.921566 4921 generic.go:334] "Generic (PLEG): container finished" podID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerID="2b8611dbc12c50a970a601b1e77eca1a06d74d3dcbd8656fea3c4e998b6be3d1" exitCode=0 Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.921795 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" event={"ID":"7dfc2c0b-f94b-4735-9ead-47ce29635737","Type":"ContainerDied","Data":"2b8611dbc12c50a970a601b1e77eca1a06d74d3dcbd8656fea3c4e998b6be3d1"} Jan 03 03:59:25 crc kubenswrapper[4921]: I0103 03:59:25.921859 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" event={"ID":"7dfc2c0b-f94b-4735-9ead-47ce29635737","Type":"ContainerStarted","Data":"ceafd8ebcb0f39592db200973b81b33b10d49d60fa1d70c7dcf53a33d5fa296c"} Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.241964 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.411774 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys\") pod \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.411815 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd\") pod \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.411841 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data\") pod \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.411869 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts\") pod \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.411928 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys\") pod \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\" (UID: \"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80\") " Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.417148 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" (UID: "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.417162 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" (UID: "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.417816 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd" (OuterVolumeSpecName: "kube-api-access-mp8dd") pod "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" (UID: "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80"). InnerVolumeSpecName "kube-api-access-mp8dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.423008 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts" (OuterVolumeSpecName: "scripts") pod "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" (UID: "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.446485 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data" (OuterVolumeSpecName: "config-data") pod "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" (UID: "34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.513577 4921 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.513622 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-kube-api-access-mp8dd\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.513636 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.513647 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.513659 4921 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.935342 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" event={"ID":"34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80","Type":"ContainerDied","Data":"970d0d7c8ab360e1d2c6cc50040c3e710bf6f4cfd39e48b0e01ca5e7b2c2bf0b"} Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.935681 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970d0d7c8ab360e1d2c6cc50040c3e710bf6f4cfd39e48b0e01ca5e7b2c2bf0b" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.935354 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6vmcw" Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.946447 4921 generic.go:334] "Generic (PLEG): container finished" podID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerID="ae27584258e3530c00fd8e2005ec193acd96010ac69bb5c104c53bd1b8610757" exitCode=0 Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.947016 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" event={"ID":"876c356f-d130-4a6f-b888-5a32ab0d83da","Type":"ContainerDied","Data":"ae27584258e3530c00fd8e2005ec193acd96010ac69bb5c104c53bd1b8610757"} Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.963698 4921 generic.go:334] "Generic (PLEG): container finished" podID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerID="60814044c4f8448d45edd1c97e241fd9bf8134192e882d185ca7752305b56f87" exitCode=0 Jan 03 03:59:26 crc kubenswrapper[4921]: I0103 03:59:26.963747 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" event={"ID":"7dfc2c0b-f94b-4735-9ead-47ce29635737","Type":"ContainerDied","Data":"60814044c4f8448d45edd1c97e241fd9bf8134192e882d185ca7752305b56f87"} Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.047238 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-79d9b68446-kxlwk"] Jan 03 03:59:27 crc kubenswrapper[4921]: E0103 03:59:27.047555 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" containerName="keystone-bootstrap" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.047573 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" containerName="keystone-bootstrap" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.047680 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" containerName="keystone-bootstrap" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.048084 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.050751 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-7c2nh" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.050762 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.050992 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.051006 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.060196 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-79d9b68446-kxlwk"] Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.222965 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-scripts\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.223061 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8x5f\" (UniqueName: \"kubernetes.io/projected/2a0b1317-5c69-485b-af20-2c12a9eb1743-kube-api-access-c8x5f\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.223138 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-fernet-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.223184 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-config-data\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.223318 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-credential-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.324518 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8x5f\" (UniqueName: \"kubernetes.io/projected/2a0b1317-5c69-485b-af20-2c12a9eb1743-kube-api-access-c8x5f\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.324609 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-fernet-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.324671 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-config-data\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.324781 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-credential-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.325762 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-scripts\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.331247 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-credential-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.331453 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-fernet-keys\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.332857 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-config-data\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.343254 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a0b1317-5c69-485b-af20-2c12a9eb1743-scripts\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.343820 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8x5f\" (UniqueName: \"kubernetes.io/projected/2a0b1317-5c69-485b-af20-2c12a9eb1743-kube-api-access-c8x5f\") pod \"keystone-79d9b68446-kxlwk\" (UID: \"2a0b1317-5c69-485b-af20-2c12a9eb1743\") " pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.413918 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.902689 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-79d9b68446-kxlwk"] Jan 03 03:59:27 crc kubenswrapper[4921]: W0103 03:59:27.909058 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a0b1317_5c69_485b_af20_2c12a9eb1743.slice/crio-2a51ec431dd35335cd70ba2543bbe7561c6092a510e2c9c28ec160160f55c609 WatchSource:0}: Error finding container 2a51ec431dd35335cd70ba2543bbe7561c6092a510e2c9c28ec160160f55c609: Status 404 returned error can't find the container with id 2a51ec431dd35335cd70ba2543bbe7561c6092a510e2c9c28ec160160f55c609 Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.975512 4921 generic.go:334] "Generic (PLEG): container finished" podID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerID="295531d23817385022f56f072cb71bd0e2bd4064221ee0f1c35eecfc340c43f1" exitCode=0 Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.975632 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" event={"ID":"7dfc2c0b-f94b-4735-9ead-47ce29635737","Type":"ContainerDied","Data":"295531d23817385022f56f072cb71bd0e2bd4064221ee0f1c35eecfc340c43f1"} Jan 03 03:59:27 crc kubenswrapper[4921]: I0103 03:59:27.977900 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" event={"ID":"2a0b1317-5c69-485b-af20-2c12a9eb1743","Type":"ContainerStarted","Data":"2a51ec431dd35335cd70ba2543bbe7561c6092a510e2c9c28ec160160f55c609"} Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.273541 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.444317 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79wsp\" (UniqueName: \"kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp\") pod \"876c356f-d130-4a6f-b888-5a32ab0d83da\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.444413 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util\") pod \"876c356f-d130-4a6f-b888-5a32ab0d83da\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.444450 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle\") pod \"876c356f-d130-4a6f-b888-5a32ab0d83da\" (UID: \"876c356f-d130-4a6f-b888-5a32ab0d83da\") " Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.445070 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle" (OuterVolumeSpecName: "bundle") pod "876c356f-d130-4a6f-b888-5a32ab0d83da" (UID: "876c356f-d130-4a6f-b888-5a32ab0d83da"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.445350 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.448721 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp" (OuterVolumeSpecName: "kube-api-access-79wsp") pod "876c356f-d130-4a6f-b888-5a32ab0d83da" (UID: "876c356f-d130-4a6f-b888-5a32ab0d83da"). InnerVolumeSpecName "kube-api-access-79wsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.457966 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util" (OuterVolumeSpecName: "util") pod "876c356f-d130-4a6f-b888-5a32ab0d83da" (UID: "876c356f-d130-4a6f-b888-5a32ab0d83da"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.546141 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79wsp\" (UniqueName: \"kubernetes.io/projected/876c356f-d130-4a6f-b888-5a32ab0d83da-kube-api-access-79wsp\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.546189 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/876c356f-d130-4a6f-b888-5a32ab0d83da-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.991116 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" event={"ID":"2a0b1317-5c69-485b-af20-2c12a9eb1743","Type":"ContainerStarted","Data":"8bccab647c37b01cc024ef9b69b620cd7ed4cc8cab170c49a6f845cb1de7a6a5"} Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.992026 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.994455 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" event={"ID":"876c356f-d130-4a6f-b888-5a32ab0d83da","Type":"ContainerDied","Data":"7ee2c8f8663c23f40b87cd8832d5e1efa2d35034c65399dea1e223dc9df327ac"} Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.994534 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ee2c8f8663c23f40b87cd8832d5e1efa2d35034c65399dea1e223dc9df327ac" Jan 03 03:59:28 crc kubenswrapper[4921]: I0103 03:59:28.994489 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.028839 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" podStartSLOduration=3.028816843 podStartE2EDuration="3.028816843s" podCreationTimestamp="2026-01-03 03:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 03:59:29.022575551 +0000 UTC m=+1104.634002395" watchObservedRunningTime="2026-01-03 03:59:29.028816843 +0000 UTC m=+1104.640243677" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.378685 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.560935 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmnsb\" (UniqueName: \"kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb\") pod \"7dfc2c0b-f94b-4735-9ead-47ce29635737\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.561129 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle\") pod \"7dfc2c0b-f94b-4735-9ead-47ce29635737\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.561184 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util\") pod \"7dfc2c0b-f94b-4735-9ead-47ce29635737\" (UID: \"7dfc2c0b-f94b-4735-9ead-47ce29635737\") " Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.562032 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle" (OuterVolumeSpecName: "bundle") pod "7dfc2c0b-f94b-4735-9ead-47ce29635737" (UID: "7dfc2c0b-f94b-4735-9ead-47ce29635737"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.567610 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb" (OuterVolumeSpecName: "kube-api-access-qmnsb") pod "7dfc2c0b-f94b-4735-9ead-47ce29635737" (UID: "7dfc2c0b-f94b-4735-9ead-47ce29635737"). InnerVolumeSpecName "kube-api-access-qmnsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.583956 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util" (OuterVolumeSpecName: "util") pod "7dfc2c0b-f94b-4735-9ead-47ce29635737" (UID: "7dfc2c0b-f94b-4735-9ead-47ce29635737"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.662664 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.662711 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7dfc2c0b-f94b-4735-9ead-47ce29635737-util\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:29 crc kubenswrapper[4921]: I0103 03:59:29.662724 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmnsb\" (UniqueName: \"kubernetes.io/projected/7dfc2c0b-f94b-4735-9ead-47ce29635737-kube-api-access-qmnsb\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:30 crc kubenswrapper[4921]: I0103 03:59:30.003635 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" event={"ID":"7dfc2c0b-f94b-4735-9ead-47ce29635737","Type":"ContainerDied","Data":"ceafd8ebcb0f39592db200973b81b33b10d49d60fa1d70c7dcf53a33d5fa296c"} Jan 03 03:59:30 crc kubenswrapper[4921]: I0103 03:59:30.004017 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceafd8ebcb0f39592db200973b81b33b10d49d60fa1d70c7dcf53a33d5fa296c" Jan 03 03:59:30 crc kubenswrapper[4921]: I0103 03:59:30.003689 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.767728 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp"] Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768577 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="util" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768594 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="util" Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768605 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="pull" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768613 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="pull" Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768621 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768627 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768641 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="util" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768646 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="util" Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768661 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="pull" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768666 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="pull" Jan 03 03:59:37 crc kubenswrapper[4921]: E0103 03:59:37.768674 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768679 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768789 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="876c356f-d130-4a6f-b888-5a32ab0d83da" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.768801 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfc2c0b-f94b-4735-9ead-47ce29635737" containerName="extract" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.769535 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.772522 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.772529 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-c5gk5" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.781295 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp"] Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.788556 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-webhook-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.788855 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnclp\" (UniqueName: \"kubernetes.io/projected/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-kube-api-access-nnclp\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.788981 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-apiservice-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.889929 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnclp\" (UniqueName: \"kubernetes.io/projected/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-kube-api-access-nnclp\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.889985 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-apiservice-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.890038 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-webhook-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.895814 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-apiservice-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.902922 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-webhook-cert\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:37 crc kubenswrapper[4921]: I0103 03:59:37.909042 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnclp\" (UniqueName: \"kubernetes.io/projected/7afe11a6-6cd5-4779-903b-fa49ae0fb5e5-kube-api-access-nnclp\") pod \"swift-operator-controller-manager-577f5cc87d-645gp\" (UID: \"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5\") " pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:38 crc kubenswrapper[4921]: I0103 03:59:38.088190 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:38 crc kubenswrapper[4921]: I0103 03:59:38.497497 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp"] Jan 03 03:59:39 crc kubenswrapper[4921]: I0103 03:59:39.065484 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" event={"ID":"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5","Type":"ContainerStarted","Data":"93f8ce0f19491620df48760ab3fb6ad535c934571ad5b0b34ae7de677182b6de"} Jan 03 03:59:42 crc kubenswrapper[4921]: I0103 03:59:42.090010 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" event={"ID":"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5","Type":"ContainerStarted","Data":"fadd4bdb0f90c74e3fd92db8e9d10b1d6a5fbae1486dc155bf9e0fa78ff38fac"} Jan 03 03:59:42 crc kubenswrapper[4921]: I0103 03:59:42.090850 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" event={"ID":"7afe11a6-6cd5-4779-903b-fa49ae0fb5e5","Type":"ContainerStarted","Data":"85ffa9111766012f6cd783a88221c67afdfb170f5a244a17fd17755c06bf1ab7"} Jan 03 03:59:42 crc kubenswrapper[4921]: I0103 03:59:42.090958 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:42 crc kubenswrapper[4921]: I0103 03:59:42.120203 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" podStartSLOduration=2.6952396480000003 podStartE2EDuration="5.120174679s" podCreationTimestamp="2026-01-03 03:59:37 +0000 UTC" firstStartedPulling="2026-01-03 03:59:38.508030765 +0000 UTC m=+1114.119457589" lastFinishedPulling="2026-01-03 03:59:40.932965796 +0000 UTC m=+1116.544392620" observedRunningTime="2026-01-03 03:59:42.111171682 +0000 UTC m=+1117.722598576" watchObservedRunningTime="2026-01-03 03:59:42.120174679 +0000 UTC m=+1117.731601543" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.527749 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7"] Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.529480 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.533605 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ht5t6" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.533789 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.547039 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7"] Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.692719 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-webhook-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.692780 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbfkf\" (UniqueName: \"kubernetes.io/projected/75035b13-a1d1-44f4-bc84-2f39ff65f476-kube-api-access-mbfkf\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.692878 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-apiservice-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.793708 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-webhook-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.793766 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbfkf\" (UniqueName: \"kubernetes.io/projected/75035b13-a1d1-44f4-bc84-2f39ff65f476-kube-api-access-mbfkf\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.793827 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-apiservice-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.799788 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-webhook-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.808591 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75035b13-a1d1-44f4-bc84-2f39ff65f476-apiservice-cert\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.816819 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbfkf\" (UniqueName: \"kubernetes.io/projected/75035b13-a1d1-44f4-bc84-2f39ff65f476-kube-api-access-mbfkf\") pod \"horizon-operator-controller-manager-6cb8f6dc8b-vtrl7\" (UID: \"75035b13-a1d1-44f4-bc84-2f39ff65f476\") " pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:44 crc kubenswrapper[4921]: I0103 03:59:44.850893 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:45 crc kubenswrapper[4921]: W0103 03:59:45.319207 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75035b13_a1d1_44f4_bc84_2f39ff65f476.slice/crio-769d58220242c14713053c75477070e595ef10c51e8b30ee961870211bd4de09 WatchSource:0}: Error finding container 769d58220242c14713053c75477070e595ef10c51e8b30ee961870211bd4de09: Status 404 returned error can't find the container with id 769d58220242c14713053c75477070e595ef10c51e8b30ee961870211bd4de09 Jan 03 03:59:45 crc kubenswrapper[4921]: I0103 03:59:45.319579 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7"] Jan 03 03:59:46 crc kubenswrapper[4921]: I0103 03:59:46.122912 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" event={"ID":"75035b13-a1d1-44f4-bc84-2f39ff65f476","Type":"ContainerStarted","Data":"769d58220242c14713053c75477070e595ef10c51e8b30ee961870211bd4de09"} Jan 03 03:59:48 crc kubenswrapper[4921]: I0103 03:59:48.092174 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-577f5cc87d-645gp" Jan 03 03:59:49 crc kubenswrapper[4921]: I0103 03:59:49.178630 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" event={"ID":"75035b13-a1d1-44f4-bc84-2f39ff65f476","Type":"ContainerStarted","Data":"d7609bd92388f64cfa3404ed1d4eab23703795459efa7b737fe592cee743bf77"} Jan 03 03:59:49 crc kubenswrapper[4921]: I0103 03:59:49.179517 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:49 crc kubenswrapper[4921]: I0103 03:59:49.179605 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" event={"ID":"75035b13-a1d1-44f4-bc84-2f39ff65f476","Type":"ContainerStarted","Data":"20021a0b53d8282043e4e5dd1f44c631606105bf3b0c574962df20a9211d5699"} Jan 03 03:59:49 crc kubenswrapper[4921]: I0103 03:59:49.195291 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" podStartSLOduration=2.399683682 podStartE2EDuration="5.195236551s" podCreationTimestamp="2026-01-03 03:59:44 +0000 UTC" firstStartedPulling="2026-01-03 03:59:45.322529809 +0000 UTC m=+1120.933956633" lastFinishedPulling="2026-01-03 03:59:48.118082638 +0000 UTC m=+1123.729509502" observedRunningTime="2026-01-03 03:59:49.192948278 +0000 UTC m=+1124.804375102" watchObservedRunningTime="2026-01-03 03:59:49.195236551 +0000 UTC m=+1124.806663395" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.408650 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.414129 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.418317 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.418724 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.427095 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.429636 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-2tjr4" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.433190 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.530385 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-lock\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.530450 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.530502 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-cache\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.530536 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.530604 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx69c\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-kube-api-access-bx69c\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.632129 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx69c\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-kube-api-access-bx69c\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.632183 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-lock\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.632218 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.632281 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-cache\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.632316 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: E0103 03:59:54.632481 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 03:59:54 crc kubenswrapper[4921]: E0103 03:59:54.632496 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Jan 03 03:59:54 crc kubenswrapper[4921]: E0103 03:59:54.632547 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift podName:5b2d14db-f843-421e-895a-92dc3986c2c7 nodeName:}" failed. No retries permitted until 2026-01-03 03:59:55.132529719 +0000 UTC m=+1130.743956553 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift") pod "swift-storage-0" (UID: "5b2d14db-f843-421e-895a-92dc3986c2c7") : configmap "swift-ring-files" not found Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.633222 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-lock\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.633563 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.637731 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b2d14db-f843-421e-895a-92dc3986c2c7-cache\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.669344 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx69c\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-kube-api-access-bx69c\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.677527 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:54 crc kubenswrapper[4921]: I0103 03:59:54.855507 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6cb8f6dc8b-vtrl7" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.030825 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-f9vjg"] Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.031919 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.033819 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.034083 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.034958 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.053417 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-f9vjg"] Jan 03 03:59:55 crc kubenswrapper[4921]: E0103 03:59:55.054237 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[dispersionconf etc-swift kube-api-access-ljf79 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[dispersionconf etc-swift kube-api-access-ljf79 ring-data-devices scripts swiftconf]: context canceled" pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" podUID="d3516129-7b18-4c16-9a21-5e7ff0328ecf" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.083119 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bckss"] Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.084142 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.091510 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bckss"] Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.100172 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-f9vjg"] Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.143948 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.143996 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.144069 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.144090 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.144112 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.144146 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljf79\" (UniqueName: \"kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.144188 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: E0103 03:59:55.144198 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 03:59:55 crc kubenswrapper[4921]: E0103 03:59:55.144216 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Jan 03 03:59:55 crc kubenswrapper[4921]: E0103 03:59:55.144607 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift podName:5b2d14db-f843-421e-895a-92dc3986c2c7 nodeName:}" failed. No retries permitted until 2026-01-03 03:59:56.14426121 +0000 UTC m=+1131.755688024 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift") pod "swift-storage-0" (UID: "5b2d14db-f843-421e-895a-92dc3986c2c7") : configmap "swift-ring-files" not found Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.218354 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.226650 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245250 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245356 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245405 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245422 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245444 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245470 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljf79\" (UniqueName: \"kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245498 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245519 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245535 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245550 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt95d\" (UniqueName: \"kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245568 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245581 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.245707 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.246174 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.246523 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.249571 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.250899 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.260032 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljf79\" (UniqueName: \"kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79\") pod \"swift-ring-rebalance-f9vjg\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.346781 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.346909 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.346941 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljf79\" (UniqueName: \"kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.346965 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.346998 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347027 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf\") pod \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\" (UID: \"d3516129-7b18-4c16-9a21-5e7ff0328ecf\") " Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347168 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347226 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347244 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347262 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt95d\" (UniqueName: \"kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347294 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347314 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.347950 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.348562 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.348666 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts" (OuterVolumeSpecName: "scripts") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.348780 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.348986 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.349000 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.350519 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.350642 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.351683 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.352157 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79" (OuterVolumeSpecName: "kube-api-access-ljf79") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "kube-api-access-ljf79". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.359551 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d3516129-7b18-4c16-9a21-5e7ff0328ecf" (UID: "d3516129-7b18-4c16-9a21-5e7ff0328ecf"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.367613 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt95d\" (UniqueName: \"kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d\") pod \"swift-ring-rebalance-bckss\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.409664 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449363 4921 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449393 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449404 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljf79\" (UniqueName: \"kubernetes.io/projected/d3516129-7b18-4c16-9a21-5e7ff0328ecf-kube-api-access-ljf79\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449413 4921 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3516129-7b18-4c16-9a21-5e7ff0328ecf-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449422 4921 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3516129-7b18-4c16-9a21-5e7ff0328ecf-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.449430 4921 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3516129-7b18-4c16-9a21-5e7ff0328ecf-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.808659 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-bckss"] Jan 03 03:59:55 crc kubenswrapper[4921]: W0103 03:59:55.812246 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37fd922c_8a9f_4e29_8af5_54e73065236b.slice/crio-ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0 WatchSource:0}: Error finding container ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0: Status 404 returned error can't find the container with id ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0 Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.987888 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.988701 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.991567 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-72sgn" Jan 03 03:59:55 crc kubenswrapper[4921]: I0103 03:59:55.996637 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.161303 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.161347 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmz9c\" (UniqueName: \"kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c\") pod \"glance-operator-index-2jrp5\" (UID: \"57b6598a-e8f1-47be-b6bc-f95bb93fe966\") " pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 03:59:56 crc kubenswrapper[4921]: E0103 03:59:56.161529 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 03:59:56 crc kubenswrapper[4921]: E0103 03:59:56.161557 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Jan 03 03:59:56 crc kubenswrapper[4921]: E0103 03:59:56.161619 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift podName:5b2d14db-f843-421e-895a-92dc3986c2c7 nodeName:}" failed. No retries permitted until 2026-01-03 03:59:58.161600321 +0000 UTC m=+1133.773027145 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift") pod "swift-storage-0" (UID: "5b2d14db-f843-421e-895a-92dc3986c2c7") : configmap "swift-ring-files" not found Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.226419 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-f9vjg" Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.226404 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" event={"ID":"37fd922c-8a9f-4e29-8af5-54e73065236b","Type":"ContainerStarted","Data":"ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0"} Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.263112 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmz9c\" (UniqueName: \"kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c\") pod \"glance-operator-index-2jrp5\" (UID: \"57b6598a-e8f1-47be-b6bc-f95bb93fe966\") " pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.276825 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-f9vjg"] Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.283689 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-f9vjg"] Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.296139 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmz9c\" (UniqueName: \"kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c\") pod \"glance-operator-index-2jrp5\" (UID: \"57b6598a-e8f1-47be-b6bc-f95bb93fe966\") " pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.353980 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.766210 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 03:59:56 crc kubenswrapper[4921]: I0103 03:59:56.895331 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3516129-7b18-4c16-9a21-5e7ff0328ecf" path="/var/lib/kubelet/pods/d3516129-7b18-4c16-9a21-5e7ff0328ecf/volumes" Jan 03 03:59:57 crc kubenswrapper[4921]: I0103 03:59:57.233061 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-2jrp5" event={"ID":"57b6598a-e8f1-47be-b6bc-f95bb93fe966","Type":"ContainerStarted","Data":"11562b195a408436b19d14835ff433473636aa19b4230fdb647da0ce2f5ea1d1"} Jan 03 03:59:58 crc kubenswrapper[4921]: I0103 03:59:58.191763 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 03:59:58 crc kubenswrapper[4921]: E0103 03:59:58.191948 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 03:59:58 crc kubenswrapper[4921]: E0103 03:59:58.192137 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Jan 03 03:59:58 crc kubenswrapper[4921]: E0103 03:59:58.192202 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift podName:5b2d14db-f843-421e-895a-92dc3986c2c7 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:02.192184623 +0000 UTC m=+1137.803611467 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift") pod "swift-storage-0" (UID: "5b2d14db-f843-421e-895a-92dc3986c2c7") : configmap "swift-ring-files" not found Jan 03 03:59:58 crc kubenswrapper[4921]: I0103 03:59:58.869517 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-79d9b68446-kxlwk" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.167260 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds"] Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.168570 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.170320 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.171743 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.176056 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds"] Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.325869 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.325941 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6h4j\" (UniqueName: \"kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.326064 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.382018 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.427669 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.427727 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6h4j\" (UniqueName: \"kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.427788 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.428787 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.444697 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.453190 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6h4j\" (UniqueName: \"kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j\") pod \"collect-profiles-29456880-q72ds\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:00 crc kubenswrapper[4921]: I0103 04:00:00.499394 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.175662 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.175728 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.214039 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-vtjh6"] Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.214822 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.219365 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-vtjh6"] Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.349803 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgq6g\" (UniqueName: \"kubernetes.io/projected/994de956-f9e4-4ba0-8b97-a9a719655c63-kube-api-access-mgq6g\") pod \"glance-operator-index-vtjh6\" (UID: \"994de956-f9e4-4ba0-8b97-a9a719655c63\") " pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.451929 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgq6g\" (UniqueName: \"kubernetes.io/projected/994de956-f9e4-4ba0-8b97-a9a719655c63-kube-api-access-mgq6g\") pod \"glance-operator-index-vtjh6\" (UID: \"994de956-f9e4-4ba0-8b97-a9a719655c63\") " pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.477388 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgq6g\" (UniqueName: \"kubernetes.io/projected/994de956-f9e4-4ba0-8b97-a9a719655c63-kube-api-access-mgq6g\") pod \"glance-operator-index-vtjh6\" (UID: \"994de956-f9e4-4ba0-8b97-a9a719655c63\") " pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:01 crc kubenswrapper[4921]: I0103 04:00:01.538046 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:02 crc kubenswrapper[4921]: I0103 04:00:02.264976 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 04:00:02 crc kubenswrapper[4921]: E0103 04:00:02.265151 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 04:00:02 crc kubenswrapper[4921]: E0103 04:00:02.265174 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Jan 03 04:00:02 crc kubenswrapper[4921]: E0103 04:00:02.265219 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift podName:5b2d14db-f843-421e-895a-92dc3986c2c7 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:10.265203921 +0000 UTC m=+1145.876630745 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift") pod "swift-storage-0" (UID: "5b2d14db-f843-421e-895a-92dc3986c2c7") : configmap "swift-ring-files" not found Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.275961 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-2jrp5" event={"ID":"57b6598a-e8f1-47be-b6bc-f95bb93fe966","Type":"ContainerStarted","Data":"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136"} Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.276012 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/glance-operator-index-2jrp5" podUID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" containerName="registry-server" containerID="cri-o://82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136" gracePeriod=2 Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.286781 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" event={"ID":"37fd922c-8a9f-4e29-8af5-54e73065236b","Type":"ContainerStarted","Data":"d28cce2701b0899fc9288c1917ee7aa7a7e09814d6d01dc67b37328216f7eaa6"} Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.299506 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds"] Jan 03 04:00:03 crc kubenswrapper[4921]: W0103 04:00:03.316137 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ae45d17_2ed2_4be3_bede_46ed1a32d110.slice/crio-d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac WatchSource:0}: Error finding container d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac: Status 404 returned error can't find the container with id d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.327378 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-2jrp5" podStartSLOduration=2.3025094729999998 podStartE2EDuration="8.327356052s" podCreationTimestamp="2026-01-03 03:59:55 +0000 UTC" firstStartedPulling="2026-01-03 03:59:56.781224711 +0000 UTC m=+1132.392651535" lastFinishedPulling="2026-01-03 04:00:02.80607126 +0000 UTC m=+1138.417498114" observedRunningTime="2026-01-03 04:00:03.306776828 +0000 UTC m=+1138.918203692" watchObservedRunningTime="2026-01-03 04:00:03.327356052 +0000 UTC m=+1138.938782886" Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.333708 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" podStartSLOduration=1.29715012 podStartE2EDuration="8.333691125s" podCreationTimestamp="2026-01-03 03:59:55 +0000 UTC" firstStartedPulling="2026-01-03 03:59:55.813932523 +0000 UTC m=+1131.425359347" lastFinishedPulling="2026-01-03 04:00:02.850473528 +0000 UTC m=+1138.461900352" observedRunningTime="2026-01-03 04:00:03.324842423 +0000 UTC m=+1138.936269257" watchObservedRunningTime="2026-01-03 04:00:03.333691125 +0000 UTC m=+1138.945117969" Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.362319 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-vtjh6"] Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.680806 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.793279 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmz9c\" (UniqueName: \"kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c\") pod \"57b6598a-e8f1-47be-b6bc-f95bb93fe966\" (UID: \"57b6598a-e8f1-47be-b6bc-f95bb93fe966\") " Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.800446 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c" (OuterVolumeSpecName: "kube-api-access-fmz9c") pod "57b6598a-e8f1-47be-b6bc-f95bb93fe966" (UID: "57b6598a-e8f1-47be-b6bc-f95bb93fe966"). InnerVolumeSpecName "kube-api-access-fmz9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:00:03 crc kubenswrapper[4921]: I0103 04:00:03.895024 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmz9c\" (UniqueName: \"kubernetes.io/projected/57b6598a-e8f1-47be-b6bc-f95bb93fe966-kube-api-access-fmz9c\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.302703 4921 generic.go:334] "Generic (PLEG): container finished" podID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" containerID="82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136" exitCode=0 Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.302786 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-2jrp5" event={"ID":"57b6598a-e8f1-47be-b6bc-f95bb93fe966","Type":"ContainerDied","Data":"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.302819 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-2jrp5" event={"ID":"57b6598a-e8f1-47be-b6bc-f95bb93fe966","Type":"ContainerDied","Data":"11562b195a408436b19d14835ff433473636aa19b4230fdb647da0ce2f5ea1d1"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.302826 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-2jrp5" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.302840 4921 scope.go:117] "RemoveContainer" containerID="82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.308454 4921 generic.go:334] "Generic (PLEG): container finished" podID="2ae45d17-2ed2-4be3-bede-46ed1a32d110" containerID="2be31a6926f50eb86a2f52222acadc1f386a8145747f0e3b3732d3163c77dbd1" exitCode=0 Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.308681 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" event={"ID":"2ae45d17-2ed2-4be3-bede-46ed1a32d110","Type":"ContainerDied","Data":"2be31a6926f50eb86a2f52222acadc1f386a8145747f0e3b3732d3163c77dbd1"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.308735 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" event={"ID":"2ae45d17-2ed2-4be3-bede-46ed1a32d110","Type":"ContainerStarted","Data":"d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.315645 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vtjh6" event={"ID":"994de956-f9e4-4ba0-8b97-a9a719655c63","Type":"ContainerStarted","Data":"6fae1e1be9bb81d1ea863a8eeec0f9691506617cd425a9589636247d47857622"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.315910 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-vtjh6" event={"ID":"994de956-f9e4-4ba0-8b97-a9a719655c63","Type":"ContainerStarted","Data":"d99140f0816ab15cd7b5985727764f94690320159865723b44c84309141932f8"} Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.335822 4921 scope.go:117] "RemoveContainer" containerID="82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136" Jan 03 04:00:04 crc kubenswrapper[4921]: E0103 04:00:04.336364 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136\": container with ID starting with 82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136 not found: ID does not exist" containerID="82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.336415 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136"} err="failed to get container status \"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136\": rpc error: code = NotFound desc = could not find container \"82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136\": container with ID starting with 82339f01772829dd09586a33bfdb84ecbd3a1a6b41b57f419dc40a6816736136 not found: ID does not exist" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.368550 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.378202 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/glance-operator-index-2jrp5"] Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.390976 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-vtjh6" podStartSLOduration=3.332260963 podStartE2EDuration="3.390954173s" podCreationTimestamp="2026-01-03 04:00:01 +0000 UTC" firstStartedPulling="2026-01-03 04:00:03.386176136 +0000 UTC m=+1138.997602970" lastFinishedPulling="2026-01-03 04:00:03.444869356 +0000 UTC m=+1139.056296180" observedRunningTime="2026-01-03 04:00:04.387836698 +0000 UTC m=+1139.999263522" watchObservedRunningTime="2026-01-03 04:00:04.390954173 +0000 UTC m=+1140.002381007" Jan 03 04:00:04 crc kubenswrapper[4921]: I0103 04:00:04.891989 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" path="/var/lib/kubelet/pods/57b6598a-e8f1-47be-b6bc-f95bb93fe966/volumes" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.191205 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6"] Jan 03 04:00:05 crc kubenswrapper[4921]: E0103 04:00:05.191497 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" containerName="registry-server" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.191509 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" containerName="registry-server" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.191653 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b6598a-e8f1-47be-b6bc-f95bb93fe966" containerName="registry-server" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.192488 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.214559 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6"] Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.430676 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sz4z\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-kube-api-access-5sz4z\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.430816 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-log-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.430859 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61567511-53a7-47d2-8b71-e57910f99a18-config-data\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.430908 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.430955 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-run-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.533175 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-log-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.533235 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61567511-53a7-47d2-8b71-e57910f99a18-config-data\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.533291 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.533325 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-run-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.533358 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sz4z\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-kube-api-access-5sz4z\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: E0103 04:00:05.533744 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 04:00:05 crc kubenswrapper[4921]: E0103 04:00:05.533762 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6: configmap "swift-ring-files" not found Jan 03 04:00:05 crc kubenswrapper[4921]: E0103 04:00:05.533800 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift podName:61567511-53a7-47d2-8b71-e57910f99a18 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:06.033787408 +0000 UTC m=+1141.645214232 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift") pod "swift-proxy-5957d6665c-x9qr6" (UID: "61567511-53a7-47d2-8b71-e57910f99a18") : configmap "swift-ring-files" not found Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.534306 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-run-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.534351 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61567511-53a7-47d2-8b71-e57910f99a18-log-httpd\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.540759 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61567511-53a7-47d2-8b71-e57910f99a18-config-data\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.554242 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sz4z\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-kube-api-access-5sz4z\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.658778 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.836912 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume\") pod \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.837305 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6h4j\" (UniqueName: \"kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j\") pod \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.837434 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume\") pod \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\" (UID: \"2ae45d17-2ed2-4be3-bede-46ed1a32d110\") " Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.838048 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume" (OuterVolumeSpecName: "config-volume") pod "2ae45d17-2ed2-4be3-bede-46ed1a32d110" (UID: "2ae45d17-2ed2-4be3-bede-46ed1a32d110"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.840928 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2ae45d17-2ed2-4be3-bede-46ed1a32d110" (UID: "2ae45d17-2ed2-4be3-bede-46ed1a32d110"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.841526 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j" (OuterVolumeSpecName: "kube-api-access-x6h4j") pod "2ae45d17-2ed2-4be3-bede-46ed1a32d110" (UID: "2ae45d17-2ed2-4be3-bede-46ed1a32d110"). InnerVolumeSpecName "kube-api-access-x6h4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.939006 4921 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2ae45d17-2ed2-4be3-bede-46ed1a32d110-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.939045 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6h4j\" (UniqueName: \"kubernetes.io/projected/2ae45d17-2ed2-4be3-bede-46ed1a32d110-kube-api-access-x6h4j\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:05 crc kubenswrapper[4921]: I0103 04:00:05.939059 4921 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2ae45d17-2ed2-4be3-bede-46ed1a32d110-config-volume\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:06 crc kubenswrapper[4921]: I0103 04:00:06.040481 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:06 crc kubenswrapper[4921]: E0103 04:00:06.040727 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 04:00:06 crc kubenswrapper[4921]: E0103 04:00:06.040751 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6: configmap "swift-ring-files" not found Jan 03 04:00:06 crc kubenswrapper[4921]: E0103 04:00:06.040801 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift podName:61567511-53a7-47d2-8b71-e57910f99a18 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:07.040783918 +0000 UTC m=+1142.652210752 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift") pod "swift-proxy-5957d6665c-x9qr6" (UID: "61567511-53a7-47d2-8b71-e57910f99a18") : configmap "swift-ring-files" not found Jan 03 04:00:06 crc kubenswrapper[4921]: I0103 04:00:06.342761 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" event={"ID":"2ae45d17-2ed2-4be3-bede-46ed1a32d110","Type":"ContainerDied","Data":"d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac"} Jan 03 04:00:06 crc kubenswrapper[4921]: I0103 04:00:06.342811 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d79898b07838754bcbf2768f4b9d52cf2e6d9c51fb98195fbf939ec8541002ac" Jan 03 04:00:06 crc kubenswrapper[4921]: I0103 04:00:06.342825 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456880-q72ds" Jan 03 04:00:07 crc kubenswrapper[4921]: I0103 04:00:07.056431 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:07 crc kubenswrapper[4921]: E0103 04:00:07.056579 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 04:00:07 crc kubenswrapper[4921]: E0103 04:00:07.056807 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6: configmap "swift-ring-files" not found Jan 03 04:00:07 crc kubenswrapper[4921]: E0103 04:00:07.056933 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift podName:61567511-53a7-47d2-8b71-e57910f99a18 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:09.056890955 +0000 UTC m=+1144.668317779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift") pod "swift-proxy-5957d6665c-x9qr6" (UID: "61567511-53a7-47d2-8b71-e57910f99a18") : configmap "swift-ring-files" not found Jan 03 04:00:09 crc kubenswrapper[4921]: I0103 04:00:09.086396 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:09 crc kubenswrapper[4921]: E0103 04:00:09.086565 4921 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Jan 03 04:00:09 crc kubenswrapper[4921]: E0103 04:00:09.086589 4921 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6: configmap "swift-ring-files" not found Jan 03 04:00:09 crc kubenswrapper[4921]: E0103 04:00:09.086653 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift podName:61567511-53a7-47d2-8b71-e57910f99a18 nodeName:}" failed. No retries permitted until 2026-01-03 04:00:13.086634434 +0000 UTC m=+1148.698061258 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift") pod "swift-proxy-5957d6665c-x9qr6" (UID: "61567511-53a7-47d2-8b71-e57910f99a18") : configmap "swift-ring-files" not found Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.305260 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.320496 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b2d14db-f843-421e-895a-92dc3986c2c7-etc-swift\") pod \"swift-storage-0\" (UID: \"5b2d14db-f843-421e-895a-92dc3986c2c7\") " pod="glance-kuttl-tests/swift-storage-0" Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.347695 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.377793 4921 generic.go:334] "Generic (PLEG): container finished" podID="37fd922c-8a9f-4e29-8af5-54e73065236b" containerID="d28cce2701b0899fc9288c1917ee7aa7a7e09814d6d01dc67b37328216f7eaa6" exitCode=0 Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.377838 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" event={"ID":"37fd922c-8a9f-4e29-8af5-54e73065236b","Type":"ContainerDied","Data":"d28cce2701b0899fc9288c1917ee7aa7a7e09814d6d01dc67b37328216f7eaa6"} Jan 03 04:00:10 crc kubenswrapper[4921]: I0103 04:00:10.979859 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.386764 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"42d3773630eb6fd972fda37bb9d53083e6b1897ea2c8ae2c1a7f7ef3fdc62682"} Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.538932 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.540436 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.577701 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.744471 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.854509 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.854610 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt95d\" (UniqueName: \"kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.854712 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.854886 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.854996 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.855071 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift\") pod \"37fd922c-8a9f-4e29-8af5-54e73065236b\" (UID: \"37fd922c-8a9f-4e29-8af5-54e73065236b\") " Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.856299 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.856890 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.864433 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d" (OuterVolumeSpecName: "kube-api-access-qt95d") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "kube-api-access-qt95d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.866904 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.874894 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts" (OuterVolumeSpecName: "scripts") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.887149 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "37fd922c-8a9f-4e29-8af5-54e73065236b" (UID: "37fd922c-8a9f-4e29-8af5-54e73065236b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.957830 4921 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.957907 4921 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/37fd922c-8a9f-4e29-8af5-54e73065236b-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.957935 4921 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.957962 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt95d\" (UniqueName: \"kubernetes.io/projected/37fd922c-8a9f-4e29-8af5-54e73065236b-kube-api-access-qt95d\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.957992 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37fd922c-8a9f-4e29-8af5-54e73065236b-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:11 crc kubenswrapper[4921]: I0103 04:00:11.958015 4921 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/37fd922c-8a9f-4e29-8af5-54e73065236b-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:12 crc kubenswrapper[4921]: I0103 04:00:12.394526 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" event={"ID":"37fd922c-8a9f-4e29-8af5-54e73065236b","Type":"ContainerDied","Data":"ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0"} Jan 03 04:00:12 crc kubenswrapper[4921]: I0103 04:00:12.395464 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec6147aacd385591fa38f2564b897e317fb4e5e488198ee3391bf12c5ddf21b0" Jan 03 04:00:12 crc kubenswrapper[4921]: I0103 04:00:12.397227 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-bckss" Jan 03 04:00:12 crc kubenswrapper[4921]: I0103 04:00:12.400293 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"c11745651ee55e3a56f273781bd38e8d2f02cb0ca22b48c6868b13607a1baf05"} Jan 03 04:00:12 crc kubenswrapper[4921]: I0103 04:00:12.437691 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-vtjh6" Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.174633 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.184681 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/61567511-53a7-47d2-8b71-e57910f99a18-etc-swift\") pod \"swift-proxy-5957d6665c-x9qr6\" (UID: \"61567511-53a7-47d2-8b71-e57910f99a18\") " pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.410975 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"58cc892c1fa88bb9b873c82bf63f667d1305aeebeb2f0f909b148eb16034556e"} Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.411033 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"0cfb1f569c2a1ce01fe9c5808f316110f2e492104a4f77bc0e7c85259dd23465"} Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.411050 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"0f74a63718adc952f078e77a455ccaf6cdbab7f735b50fead1a4e82ca186738c"} Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.431759 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:13 crc kubenswrapper[4921]: I0103 04:00:13.886561 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6"] Jan 03 04:00:16 crc kubenswrapper[4921]: I0103 04:00:16.444663 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" event={"ID":"61567511-53a7-47d2-8b71-e57910f99a18","Type":"ContainerStarted","Data":"9386cdcf4ce26653048d2422efd4e805e4479e781c2d5f0f6750a14371ef6301"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.456412 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" event={"ID":"61567511-53a7-47d2-8b71-e57910f99a18","Type":"ContainerStarted","Data":"5ab535289ade92b36fedd507128cbc3ac5677c2a0580d54a649f1123f32effef"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.456763 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" event={"ID":"61567511-53a7-47d2-8b71-e57910f99a18","Type":"ContainerStarted","Data":"1498f32c43bec5725d0e7fdcf27a77595a93682d60454103f7d7e6c9af096c45"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.456785 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.456798 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.502913 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" podStartSLOduration=12.502893023 podStartE2EDuration="12.502893023s" podCreationTimestamp="2026-01-03 04:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:00:17.499742337 +0000 UTC m=+1153.111169181" watchObservedRunningTime="2026-01-03 04:00:17.502893023 +0000 UTC m=+1153.114319857" Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.529581 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"b3e89660e2e7808e2260b6338e6142de29226eb531338a22d5195f8a120738c7"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.529639 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"b6852aa18acae65ee464b816861966f58c6f12de4ff7e0c7598c2f110c36416c"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.529652 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"c6610b50606d96e2cb3829d698054379f31ef368be5e9277142af150b5fb0bc5"} Jan 03 04:00:17 crc kubenswrapper[4921]: I0103 04:00:17.529664 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"840f891d41b5f49762a5d4bf73999174c9c11d9c4659c90f072a1b80f9191195"} Jan 03 04:00:18 crc kubenswrapper[4921]: I0103 04:00:18.545980 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"78fc15bd922216711220c969ca53463777b95c57a08507e4369547d0f77e6701"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.237384 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67"] Jan 03 04:00:19 crc kubenswrapper[4921]: E0103 04:00:19.237873 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37fd922c-8a9f-4e29-8af5-54e73065236b" containerName="swift-ring-rebalance" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.237898 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="37fd922c-8a9f-4e29-8af5-54e73065236b" containerName="swift-ring-rebalance" Jan 03 04:00:19 crc kubenswrapper[4921]: E0103 04:00:19.237926 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae45d17-2ed2-4be3-bede-46ed1a32d110" containerName="collect-profiles" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.237935 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae45d17-2ed2-4be3-bede-46ed1a32d110" containerName="collect-profiles" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.238129 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae45d17-2ed2-4be3-bede-46ed1a32d110" containerName="collect-profiles" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.238150 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="37fd922c-8a9f-4e29-8af5-54e73065236b" containerName="swift-ring-rebalance" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.239437 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.241521 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8jbqk" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.244981 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67"] Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.272677 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdflk\" (UniqueName: \"kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.272813 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.273070 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.373829 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.373899 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdflk\" (UniqueName: \"kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.373919 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.374309 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.374382 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.397044 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdflk\" (UniqueName: \"kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk\") pod \"1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.557062 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"b4bda10ac309a6167d6526c413fdcb2de7291435e2fba57c007d932e4985f09d"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.557108 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"cee1500dd0b61dddc3bcc47b84b9a0b935a03228336eae7125bc1e579f989204"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.557121 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"e4a27b8c2cdeeba647895f1a662649f97dbb5a49ca9ef59f25f254b33dfbb02d"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.557135 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"65d846861913bfb0e12f714d760c1a6b896a714b22440e94d019119a9ea204a2"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.557148 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"318c3995089699a72d3c8a3971ae58c432154e7b0c3e3d83096effb146a81b94"} Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.599216 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:19 crc kubenswrapper[4921]: I0103 04:00:19.884131 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67"] Jan 03 04:00:19 crc kubenswrapper[4921]: W0103 04:00:19.889124 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70ec3f78_d438_442a_8378_d23a76219ca0.slice/crio-4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9 WatchSource:0}: Error finding container 4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9: Status 404 returned error can't find the container with id 4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9 Jan 03 04:00:20 crc kubenswrapper[4921]: I0103 04:00:20.586155 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"5b2d14db-f843-421e-895a-92dc3986c2c7","Type":"ContainerStarted","Data":"61b762c21890df5779139fcf4233b9bddf399982a9d9ef75c3f36c8755046d62"} Jan 03 04:00:20 crc kubenswrapper[4921]: I0103 04:00:20.598479 4921 generic.go:334] "Generic (PLEG): container finished" podID="70ec3f78-d438-442a-8378-d23a76219ca0" containerID="51f02e8168c926e7ec30f76c6eb8572b5f508c44563dfef2bea1329c4405683a" exitCode=0 Jan 03 04:00:20 crc kubenswrapper[4921]: I0103 04:00:20.598550 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" event={"ID":"70ec3f78-d438-442a-8378-d23a76219ca0","Type":"ContainerDied","Data":"51f02e8168c926e7ec30f76c6eb8572b5f508c44563dfef2bea1329c4405683a"} Jan 03 04:00:20 crc kubenswrapper[4921]: I0103 04:00:20.598618 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" event={"ID":"70ec3f78-d438-442a-8378-d23a76219ca0","Type":"ContainerStarted","Data":"4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9"} Jan 03 04:00:20 crc kubenswrapper[4921]: I0103 04:00:20.641894 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=20.404251573 podStartE2EDuration="27.641868105s" podCreationTimestamp="2026-01-03 03:59:53 +0000 UTC" firstStartedPulling="2026-01-03 04:00:10.998750705 +0000 UTC m=+1146.610177549" lastFinishedPulling="2026-01-03 04:00:18.236367257 +0000 UTC m=+1153.847794081" observedRunningTime="2026-01-03 04:00:20.636140938 +0000 UTC m=+1156.247567802" watchObservedRunningTime="2026-01-03 04:00:20.641868105 +0000 UTC m=+1156.253294959" Jan 03 04:00:21 crc kubenswrapper[4921]: I0103 04:00:21.608350 4921 generic.go:334] "Generic (PLEG): container finished" podID="70ec3f78-d438-442a-8378-d23a76219ca0" containerID="743fcded6c116d0d804a46a2c59737ec3e2b7aaf9e47cfef215427a6440b6afe" exitCode=0 Jan 03 04:00:21 crc kubenswrapper[4921]: I0103 04:00:21.608392 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" event={"ID":"70ec3f78-d438-442a-8378-d23a76219ca0","Type":"ContainerDied","Data":"743fcded6c116d0d804a46a2c59737ec3e2b7aaf9e47cfef215427a6440b6afe"} Jan 03 04:00:22 crc kubenswrapper[4921]: I0103 04:00:22.619674 4921 generic.go:334] "Generic (PLEG): container finished" podID="70ec3f78-d438-442a-8378-d23a76219ca0" containerID="68d3f95d05e6fa41a3b44b48c4b725c900a8f87ca38e37e4a49fd15810ca35ab" exitCode=0 Jan 03 04:00:22 crc kubenswrapper[4921]: I0103 04:00:22.619751 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" event={"ID":"70ec3f78-d438-442a-8378-d23a76219ca0","Type":"ContainerDied","Data":"68d3f95d05e6fa41a3b44b48c4b725c900a8f87ca38e37e4a49fd15810ca35ab"} Jan 03 04:00:23 crc kubenswrapper[4921]: I0103 04:00:23.434924 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:23 crc kubenswrapper[4921]: I0103 04:00:23.438631 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.035417 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.149569 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdflk\" (UniqueName: \"kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk\") pod \"70ec3f78-d438-442a-8378-d23a76219ca0\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.149635 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util\") pod \"70ec3f78-d438-442a-8378-d23a76219ca0\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.149728 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle\") pod \"70ec3f78-d438-442a-8378-d23a76219ca0\" (UID: \"70ec3f78-d438-442a-8378-d23a76219ca0\") " Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.150637 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle" (OuterVolumeSpecName: "bundle") pod "70ec3f78-d438-442a-8378-d23a76219ca0" (UID: "70ec3f78-d438-442a-8378-d23a76219ca0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.156371 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk" (OuterVolumeSpecName: "kube-api-access-jdflk") pod "70ec3f78-d438-442a-8378-d23a76219ca0" (UID: "70ec3f78-d438-442a-8378-d23a76219ca0"). InnerVolumeSpecName "kube-api-access-jdflk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.164673 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util" (OuterVolumeSpecName: "util") pod "70ec3f78-d438-442a-8378-d23a76219ca0" (UID: "70ec3f78-d438-442a-8378-d23a76219ca0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.251476 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdflk\" (UniqueName: \"kubernetes.io/projected/70ec3f78-d438-442a-8378-d23a76219ca0-kube-api-access-jdflk\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.251514 4921 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-util\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.251528 4921 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70ec3f78-d438-442a-8378-d23a76219ca0-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.708257 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" event={"ID":"70ec3f78-d438-442a-8378-d23a76219ca0","Type":"ContainerDied","Data":"4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9"} Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.711425 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ce4aee2dd2b988e9b2c381c832829c1d28fa31200f1c799ca4af1d4a4fab0e9" Jan 03 04:00:24 crc kubenswrapper[4921]: I0103 04:00:24.712185 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67" Jan 03 04:00:31 crc kubenswrapper[4921]: I0103 04:00:31.176551 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:00:31 crc kubenswrapper[4921]: I0103 04:00:31.177062 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.026059 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7"] Jan 03 04:00:35 crc kubenswrapper[4921]: E0103 04:00:35.027507 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="pull" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.027591 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="pull" Jan 03 04:00:35 crc kubenswrapper[4921]: E0103 04:00:35.027656 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="extract" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.027722 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="extract" Jan 03 04:00:35 crc kubenswrapper[4921]: E0103 04:00:35.027794 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="util" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.027850 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="util" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.028041 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ec3f78-d438-442a-8378-d23a76219ca0" containerName="extract" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.028790 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.031835 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.032304 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-d25qm" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.051741 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7"] Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.158757 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-webhook-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.158798 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-apiservice-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.159078 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhbz\" (UniqueName: \"kubernetes.io/projected/771c903b-35e6-4f43-9b20-ab43e6726425-kube-api-access-tfhbz\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.259887 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-webhook-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.259928 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-apiservice-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.260006 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhbz\" (UniqueName: \"kubernetes.io/projected/771c903b-35e6-4f43-9b20-ab43e6726425-kube-api-access-tfhbz\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.267365 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-apiservice-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.278799 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhbz\" (UniqueName: \"kubernetes.io/projected/771c903b-35e6-4f43-9b20-ab43e6726425-kube-api-access-tfhbz\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.281142 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/771c903b-35e6-4f43-9b20-ab43e6726425-webhook-cert\") pod \"glance-operator-controller-manager-b8df6954c-rpzb7\" (UID: \"771c903b-35e6-4f43-9b20-ab43e6726425\") " pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.347989 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:35 crc kubenswrapper[4921]: I0103 04:00:35.816385 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7"] Jan 03 04:00:36 crc kubenswrapper[4921]: I0103 04:00:36.807574 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" event={"ID":"771c903b-35e6-4f43-9b20-ab43e6726425","Type":"ContainerStarted","Data":"fac5d689c9d5773613eea4dc52f7601b38d06537c6f08bd240725360e201dc10"} Jan 03 04:00:37 crc kubenswrapper[4921]: I0103 04:00:37.822019 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" event={"ID":"771c903b-35e6-4f43-9b20-ab43e6726425","Type":"ContainerStarted","Data":"882e424e2921696733c7c4be08a04af20897fa511e487bb647f280a9ace91ebb"} Jan 03 04:00:38 crc kubenswrapper[4921]: I0103 04:00:38.832726 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" event={"ID":"771c903b-35e6-4f43-9b20-ab43e6726425","Type":"ContainerStarted","Data":"3b63ad6c508c1fcea743523e8ad85494ec3ecb1263670fd5518cb1ab76d5b809"} Jan 03 04:00:38 crc kubenswrapper[4921]: I0103 04:00:38.833306 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:38 crc kubenswrapper[4921]: I0103 04:00:38.858152 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" podStartSLOduration=1.57810399 podStartE2EDuration="3.858122858s" podCreationTimestamp="2026-01-03 04:00:35 +0000 UTC" firstStartedPulling="2026-01-03 04:00:35.830407573 +0000 UTC m=+1171.441834427" lastFinishedPulling="2026-01-03 04:00:38.110426471 +0000 UTC m=+1173.721853295" observedRunningTime="2026-01-03 04:00:38.851084387 +0000 UTC m=+1174.462511251" watchObservedRunningTime="2026-01-03 04:00:38.858122858 +0000 UTC m=+1174.469549692" Jan 03 04:00:45 crc kubenswrapper[4921]: I0103 04:00:45.353179 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-b8df6954c-rpzb7" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.810087 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.811472 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.813864 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-sh9p2" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.814020 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.814635 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.815965 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.822538 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.842077 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-b47hv"] Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.843143 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.854526 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-b47hv"] Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.979499 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config-secret\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.979586 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.979604 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpdwp\" (UniqueName: \"kubernetes.io/projected/63d57045-4c96-4700-8450-8c3bdf348bd3-kube-api-access-xpdwp\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.979645 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw58q\" (UniqueName: \"kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q\") pod \"glance-db-create-b47hv\" (UID: \"0409c8d5-d153-4227-99c2-45414648ba2c\") " pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:48 crc kubenswrapper[4921]: I0103 04:00:48.979661 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-scripts\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.081460 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config-secret\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.081659 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.081705 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpdwp\" (UniqueName: \"kubernetes.io/projected/63d57045-4c96-4700-8450-8c3bdf348bd3-kube-api-access-xpdwp\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.081797 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw58q\" (UniqueName: \"kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q\") pod \"glance-db-create-b47hv\" (UID: \"0409c8d5-d153-4227-99c2-45414648ba2c\") " pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.081851 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-scripts\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.082708 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.083749 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-scripts\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.088089 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/63d57045-4c96-4700-8450-8c3bdf348bd3-openstack-config-secret\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.104327 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpdwp\" (UniqueName: \"kubernetes.io/projected/63d57045-4c96-4700-8450-8c3bdf348bd3-kube-api-access-xpdwp\") pod \"openstackclient\" (UID: \"63d57045-4c96-4700-8450-8c3bdf348bd3\") " pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.104563 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw58q\" (UniqueName: \"kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q\") pod \"glance-db-create-b47hv\" (UID: \"0409c8d5-d153-4227-99c2-45414648ba2c\") " pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.131642 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.162946 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.398865 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.435759 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-b47hv"] Jan 03 04:00:49 crc kubenswrapper[4921]: W0103 04:00:49.443719 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0409c8d5_d153_4227_99c2_45414648ba2c.slice/crio-4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d WatchSource:0}: Error finding container 4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d: Status 404 returned error can't find the container with id 4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.919739 4921 generic.go:334] "Generic (PLEG): container finished" podID="0409c8d5-d153-4227-99c2-45414648ba2c" containerID="a35139db0766dc14a29869569638c50cf2adba17eb3a5cbee63ca5128426f596" exitCode=0 Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.919797 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-b47hv" event={"ID":"0409c8d5-d153-4227-99c2-45414648ba2c","Type":"ContainerDied","Data":"a35139db0766dc14a29869569638c50cf2adba17eb3a5cbee63ca5128426f596"} Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.919847 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-b47hv" event={"ID":"0409c8d5-d153-4227-99c2-45414648ba2c","Type":"ContainerStarted","Data":"4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d"} Jan 03 04:00:49 crc kubenswrapper[4921]: I0103 04:00:49.921694 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"63d57045-4c96-4700-8450-8c3bdf348bd3","Type":"ContainerStarted","Data":"345c73875c93d80178b2ec3ffa3ac0d9d817306906224da3236943b616f61157"} Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.234839 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.318620 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw58q\" (UniqueName: \"kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q\") pod \"0409c8d5-d153-4227-99c2-45414648ba2c\" (UID: \"0409c8d5-d153-4227-99c2-45414648ba2c\") " Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.328767 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q" (OuterVolumeSpecName: "kube-api-access-kw58q") pod "0409c8d5-d153-4227-99c2-45414648ba2c" (UID: "0409c8d5-d153-4227-99c2-45414648ba2c"). InnerVolumeSpecName "kube-api-access-kw58q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.421308 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw58q\" (UniqueName: \"kubernetes.io/projected/0409c8d5-d153-4227-99c2-45414648ba2c-kube-api-access-kw58q\") on node \"crc\" DevicePath \"\"" Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.940507 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-b47hv" event={"ID":"0409c8d5-d153-4227-99c2-45414648ba2c","Type":"ContainerDied","Data":"4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d"} Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.940552 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bec2ffe6535cc0abaff6c4f7b26ee8a4cf24c85bde535ba43e2748da857325d" Jan 03 04:00:51 crc kubenswrapper[4921]: I0103 04:00:51.940614 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-b47hv" Jan 03 04:00:56 crc kubenswrapper[4921]: I0103 04:00:56.981061 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"63d57045-4c96-4700-8450-8c3bdf348bd3","Type":"ContainerStarted","Data":"8ed6b01efcb9a20d614e868cc8db26d08eff963d56962a1b1759591bd611fde5"} Jan 03 04:00:57 crc kubenswrapper[4921]: I0103 04:00:57.000131 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=1.690342596 podStartE2EDuration="9.000113254s" podCreationTimestamp="2026-01-03 04:00:48 +0000 UTC" firstStartedPulling="2026-01-03 04:00:49.406243021 +0000 UTC m=+1185.017669845" lastFinishedPulling="2026-01-03 04:00:56.716013679 +0000 UTC m=+1192.327440503" observedRunningTime="2026-01-03 04:00:56.998086074 +0000 UTC m=+1192.609512918" watchObservedRunningTime="2026-01-03 04:00:57.000113254 +0000 UTC m=+1192.611540078" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.851829 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-a6c3-account-create-9cb66"] Jan 03 04:00:58 crc kubenswrapper[4921]: E0103 04:00:58.852480 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0409c8d5-d153-4227-99c2-45414648ba2c" containerName="mariadb-database-create" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.852496 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="0409c8d5-d153-4227-99c2-45414648ba2c" containerName="mariadb-database-create" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.852685 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="0409c8d5-d153-4227-99c2-45414648ba2c" containerName="mariadb-database-create" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.853256 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.855658 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.867408 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a6c3-account-create-9cb66"] Jan 03 04:00:58 crc kubenswrapper[4921]: I0103 04:00:58.941370 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r68cx\" (UniqueName: \"kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx\") pod \"glance-a6c3-account-create-9cb66\" (UID: \"daf030d9-6b0b-4992-b355-68c358153bce\") " pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:00:59 crc kubenswrapper[4921]: I0103 04:00:59.043618 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r68cx\" (UniqueName: \"kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx\") pod \"glance-a6c3-account-create-9cb66\" (UID: \"daf030d9-6b0b-4992-b355-68c358153bce\") " pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:00:59 crc kubenswrapper[4921]: I0103 04:00:59.065489 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r68cx\" (UniqueName: \"kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx\") pod \"glance-a6c3-account-create-9cb66\" (UID: \"daf030d9-6b0b-4992-b355-68c358153bce\") " pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:00:59 crc kubenswrapper[4921]: I0103 04:00:59.181752 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:00:59 crc kubenswrapper[4921]: I0103 04:00:59.640613 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-a6c3-account-create-9cb66"] Jan 03 04:00:59 crc kubenswrapper[4921]: W0103 04:00:59.657699 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf030d9_6b0b_4992_b355_68c358153bce.slice/crio-f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894 WatchSource:0}: Error finding container f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894: Status 404 returned error can't find the container with id f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894 Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.007096 4921 generic.go:334] "Generic (PLEG): container finished" podID="daf030d9-6b0b-4992-b355-68c358153bce" containerID="62da10a9f4cd1fb2a083feaa6a9a63b9f43e5c16064bd03065dfb82818a1502b" exitCode=0 Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.007164 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" event={"ID":"daf030d9-6b0b-4992-b355-68c358153bce","Type":"ContainerDied","Data":"62da10a9f4cd1fb2a083feaa6a9a63b9f43e5c16064bd03065dfb82818a1502b"} Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.007233 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" event={"ID":"daf030d9-6b0b-4992-b355-68c358153bce","Type":"ContainerStarted","Data":"f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894"} Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.135381 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-cron-29456881-9mwqb"] Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.137038 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.144743 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29456881-9mwqb"] Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.263809 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvpmh\" (UniqueName: \"kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.263867 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.263941 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.366347 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvpmh\" (UniqueName: \"kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.366756 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.366996 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.375769 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.387960 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.398394 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvpmh\" (UniqueName: \"kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh\") pod \"keystone-cron-29456881-9mwqb\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.460524 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:00 crc kubenswrapper[4921]: I0103 04:01:00.733497 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-cron-29456881-9mwqb"] Jan 03 04:01:00 crc kubenswrapper[4921]: W0103 04:01:00.738906 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod805af972_9220_4b06_a94a_bca97464b177.slice/crio-54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2 WatchSource:0}: Error finding container 54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2: Status 404 returned error can't find the container with id 54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2 Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.022164 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" event={"ID":"805af972-9220-4b06-a94a-bca97464b177","Type":"ContainerStarted","Data":"5be1701a87940c123a306ad1896c0bf065a3f929da311d31a21fb0728e2ae47e"} Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.022547 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" event={"ID":"805af972-9220-4b06-a94a-bca97464b177","Type":"ContainerStarted","Data":"54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2"} Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.045951 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" podStartSLOduration=1.045929919 podStartE2EDuration="1.045929919s" podCreationTimestamp="2026-01-03 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:01:01.036692251 +0000 UTC m=+1196.648119085" watchObservedRunningTime="2026-01-03 04:01:01.045929919 +0000 UTC m=+1196.657356743" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.183324 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.183377 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.183418 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.184198 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.184296 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4" gracePeriod=600 Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.262207 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.387736 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r68cx\" (UniqueName: \"kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx\") pod \"daf030d9-6b0b-4992-b355-68c358153bce\" (UID: \"daf030d9-6b0b-4992-b355-68c358153bce\") " Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.397521 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx" (OuterVolumeSpecName: "kube-api-access-r68cx") pod "daf030d9-6b0b-4992-b355-68c358153bce" (UID: "daf030d9-6b0b-4992-b355-68c358153bce"). InnerVolumeSpecName "kube-api-access-r68cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:01:01 crc kubenswrapper[4921]: I0103 04:01:01.490049 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r68cx\" (UniqueName: \"kubernetes.io/projected/daf030d9-6b0b-4992-b355-68c358153bce-kube-api-access-r68cx\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.034716 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4" exitCode=0 Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.035037 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4"} Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.035328 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3"} Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.035353 4921 scope.go:117] "RemoveContainer" containerID="f402fd4335520efa0227a9a2b36ded51c7415372b9fd0f1da48b38c17362ecdd" Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.039252 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" event={"ID":"daf030d9-6b0b-4992-b355-68c358153bce","Type":"ContainerDied","Data":"f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894"} Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.039321 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-a6c3-account-create-9cb66" Jan 03 04:01:02 crc kubenswrapper[4921]: I0103 04:01:02.039343 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1d10ac311b32f4fe78207e0a44f6c65f6f62526da8855e81b2ee0234860a894" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.057165 4921 generic.go:334] "Generic (PLEG): container finished" podID="805af972-9220-4b06-a94a-bca97464b177" containerID="5be1701a87940c123a306ad1896c0bf065a3f929da311d31a21fb0728e2ae47e" exitCode=0 Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.057334 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" event={"ID":"805af972-9220-4b06-a94a-bca97464b177","Type":"ContainerDied","Data":"5be1701a87940c123a306ad1896c0bf065a3f929da311d31a21fb0728e2ae47e"} Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.903823 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-6c824"] Jan 03 04:01:03 crc kubenswrapper[4921]: E0103 04:01:03.904220 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf030d9-6b0b-4992-b355-68c358153bce" containerName="mariadb-account-create" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.904242 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf030d9-6b0b-4992-b355-68c358153bce" containerName="mariadb-account-create" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.904582 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf030d9-6b0b-4992-b355-68c358153bce" containerName="mariadb-account-create" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.905328 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.907944 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.909944 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-ddhvq" Jan 03 04:01:03 crc kubenswrapper[4921]: I0103 04:01:03.921910 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6c824"] Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.026997 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.027050 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.027132 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wb9n\" (UniqueName: \"kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.129014 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.130486 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.130614 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wb9n\" (UniqueName: \"kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.137484 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.150009 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.167248 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wb9n\" (UniqueName: \"kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n\") pod \"glance-db-sync-6c824\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.236130 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.343281 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.435845 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data\") pod \"805af972-9220-4b06-a94a-bca97464b177\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.435941 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvpmh\" (UniqueName: \"kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh\") pod \"805af972-9220-4b06-a94a-bca97464b177\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.435966 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys\") pod \"805af972-9220-4b06-a94a-bca97464b177\" (UID: \"805af972-9220-4b06-a94a-bca97464b177\") " Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.440395 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh" (OuterVolumeSpecName: "kube-api-access-mvpmh") pod "805af972-9220-4b06-a94a-bca97464b177" (UID: "805af972-9220-4b06-a94a-bca97464b177"). InnerVolumeSpecName "kube-api-access-mvpmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.440918 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "805af972-9220-4b06-a94a-bca97464b177" (UID: "805af972-9220-4b06-a94a-bca97464b177"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.458542 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6c824"] Jan 03 04:01:04 crc kubenswrapper[4921]: W0103 04:01:04.459201 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57fda7f2_4f05_4abb_80c3_686dda011b05.slice/crio-27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610 WatchSource:0}: Error finding container 27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610: Status 404 returned error can't find the container with id 27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610 Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.472754 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data" (OuterVolumeSpecName: "config-data") pod "805af972-9220-4b06-a94a-bca97464b177" (UID: "805af972-9220-4b06-a94a-bca97464b177"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.538741 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.538780 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvpmh\" (UniqueName: \"kubernetes.io/projected/805af972-9220-4b06-a94a-bca97464b177-kube-api-access-mvpmh\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:04 crc kubenswrapper[4921]: I0103 04:01:04.538790 4921 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/805af972-9220-4b06-a94a-bca97464b177-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:05 crc kubenswrapper[4921]: I0103 04:01:05.073969 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" event={"ID":"805af972-9220-4b06-a94a-bca97464b177","Type":"ContainerDied","Data":"54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2"} Jan 03 04:01:05 crc kubenswrapper[4921]: I0103 04:01:05.074019 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54230fd3a3be611e054ba7cedfa89ff34728c53bc6cc3c5752fc7f292e3641f2" Jan 03 04:01:05 crc kubenswrapper[4921]: I0103 04:01:05.074019 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-cron-29456881-9mwqb" Jan 03 04:01:05 crc kubenswrapper[4921]: I0103 04:01:05.076714 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6c824" event={"ID":"57fda7f2-4f05-4abb-80c3-686dda011b05","Type":"ContainerStarted","Data":"27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610"} Jan 03 04:01:18 crc kubenswrapper[4921]: I0103 04:01:18.182709 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6c824" event={"ID":"57fda7f2-4f05-4abb-80c3-686dda011b05","Type":"ContainerStarted","Data":"1a1d9d007caf27d0bf2c979d1de8879cab94c0ec124f85a95b0d026333bda6b9"} Jan 03 04:01:18 crc kubenswrapper[4921]: I0103 04:01:18.209251 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-6c824" podStartSLOduration=2.7620694180000003 podStartE2EDuration="15.209233887s" podCreationTimestamp="2026-01-03 04:01:03 +0000 UTC" firstStartedPulling="2026-01-03 04:01:04.462724348 +0000 UTC m=+1200.074151172" lastFinishedPulling="2026-01-03 04:01:16.909888777 +0000 UTC m=+1212.521315641" observedRunningTime="2026-01-03 04:01:18.202852316 +0000 UTC m=+1213.814279150" watchObservedRunningTime="2026-01-03 04:01:18.209233887 +0000 UTC m=+1213.820660711" Jan 03 04:01:24 crc kubenswrapper[4921]: I0103 04:01:24.231101 4921 generic.go:334] "Generic (PLEG): container finished" podID="57fda7f2-4f05-4abb-80c3-686dda011b05" containerID="1a1d9d007caf27d0bf2c979d1de8879cab94c0ec124f85a95b0d026333bda6b9" exitCode=0 Jan 03 04:01:24 crc kubenswrapper[4921]: I0103 04:01:24.231434 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6c824" event={"ID":"57fda7f2-4f05-4abb-80c3-686dda011b05","Type":"ContainerDied","Data":"1a1d9d007caf27d0bf2c979d1de8879cab94c0ec124f85a95b0d026333bda6b9"} Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.521875 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.646785 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data\") pod \"57fda7f2-4f05-4abb-80c3-686dda011b05\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.647117 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data\") pod \"57fda7f2-4f05-4abb-80c3-686dda011b05\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.647240 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wb9n\" (UniqueName: \"kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n\") pod \"57fda7f2-4f05-4abb-80c3-686dda011b05\" (UID: \"57fda7f2-4f05-4abb-80c3-686dda011b05\") " Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.652073 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n" (OuterVolumeSpecName: "kube-api-access-7wb9n") pod "57fda7f2-4f05-4abb-80c3-686dda011b05" (UID: "57fda7f2-4f05-4abb-80c3-686dda011b05"). InnerVolumeSpecName "kube-api-access-7wb9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.653518 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "57fda7f2-4f05-4abb-80c3-686dda011b05" (UID: "57fda7f2-4f05-4abb-80c3-686dda011b05"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.698797 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data" (OuterVolumeSpecName: "config-data") pod "57fda7f2-4f05-4abb-80c3-686dda011b05" (UID: "57fda7f2-4f05-4abb-80c3-686dda011b05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.749813 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wb9n\" (UniqueName: \"kubernetes.io/projected/57fda7f2-4f05-4abb-80c3-686dda011b05-kube-api-access-7wb9n\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.749857 4921 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:25 crc kubenswrapper[4921]: I0103 04:01:25.749870 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57fda7f2-4f05-4abb-80c3-686dda011b05-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:26 crc kubenswrapper[4921]: I0103 04:01:26.249179 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-6c824" event={"ID":"57fda7f2-4f05-4abb-80c3-686dda011b05","Type":"ContainerDied","Data":"27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610"} Jan 03 04:01:26 crc kubenswrapper[4921]: I0103 04:01:26.249217 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27d436984d7cbc0e4cfdeb069e010c00909f2c25f9da011cccdf8f3e0ff61610" Jan 03 04:01:26 crc kubenswrapper[4921]: I0103 04:01:26.249265 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-6c824" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.690703 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:01:27 crc kubenswrapper[4921]: E0103 04:01:27.691259 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="805af972-9220-4b06-a94a-bca97464b177" containerName="keystone-cron" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.691298 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="805af972-9220-4b06-a94a-bca97464b177" containerName="keystone-cron" Jan 03 04:01:27 crc kubenswrapper[4921]: E0103 04:01:27.691313 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57fda7f2-4f05-4abb-80c3-686dda011b05" containerName="glance-db-sync" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.691320 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="57fda7f2-4f05-4abb-80c3-686dda011b05" containerName="glance-db-sync" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.691505 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="57fda7f2-4f05-4abb-80c3-686dda011b05" containerName="glance-db-sync" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.691524 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="805af972-9220-4b06-a94a-bca97464b177" containerName="keystone-cron" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.692404 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.694425 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.694452 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.694708 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-ddhvq" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.705902 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.708087 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.722087 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.733294 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879484 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqpt8\" (UniqueName: \"kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879540 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879570 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879604 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879626 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtq57\" (UniqueName: \"kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879670 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879692 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879720 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879745 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879767 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.879958 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880002 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880050 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880090 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880130 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880174 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880199 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880329 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880365 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880393 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880430 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880451 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880473 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880497 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880518 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880542 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880564 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.880593 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981725 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981771 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981806 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981826 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981852 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981872 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981894 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981927 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981950 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981968 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981968 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.981986 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982037 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982043 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982062 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982095 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982127 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982144 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982196 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982199 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982213 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982171 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqpt8\" (UniqueName: \"kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982167 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982163 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982177 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982351 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982389 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982411 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982450 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982473 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtq57\" (UniqueName: \"kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982476 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982538 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982597 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982501 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982689 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982750 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982769 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.982815 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983042 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983465 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983550 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983583 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983656 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983677 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983709 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983834 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.983863 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.984108 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.984156 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.984183 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.988047 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.989028 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.989892 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:27 crc kubenswrapper[4921]: I0103 04:01:27.992395 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.005262 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqpt8\" (UniqueName: \"kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.005659 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.012774 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.014445 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.016386 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtq57\" (UniqueName: \"kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57\") pod \"glance-default-single-1\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.052919 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.312726 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.325825 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.614315 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:28 crc kubenswrapper[4921]: W0103 04:01:28.658952 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0fcf55e_ea3e_4b3a_b06f_dfd649e8c0e4.slice/crio-ee5fe17bad747878c86f86cd362f85f4b3f62e11f106bdfee3a869de56513618 WatchSource:0}: Error finding container ee5fe17bad747878c86f86cd362f85f4b3f62e11f106bdfee3a869de56513618: Status 404 returned error can't find the container with id ee5fe17bad747878c86f86cd362f85f4b3f62e11f106bdfee3a869de56513618 Jan 03 04:01:28 crc kubenswrapper[4921]: I0103 04:01:28.660041 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.279621 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerStarted","Data":"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.283365 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerStarted","Data":"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.283643 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerStarted","Data":"ee5fe17bad747878c86f86cd362f85f4b3f62e11f106bdfee3a869de56513618"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.286394 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerStarted","Data":"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.286465 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerStarted","Data":"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.286485 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerStarted","Data":"d33b2e77eb686304f785f2fae83eb1f1e05041fa247ee11ee8e53ed49a6fa1ce"} Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.318622 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.318601072 podStartE2EDuration="2.318601072s" podCreationTimestamp="2026-01-03 04:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:01:29.310890081 +0000 UTC m=+1224.922316975" watchObservedRunningTime="2026-01-03 04:01:29.318601072 +0000 UTC m=+1224.930027906" Jan 03 04:01:29 crc kubenswrapper[4921]: I0103 04:01:29.345901 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.345886872 podStartE2EDuration="3.345886872s" podCreationTimestamp="2026-01-03 04:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:01:29.344977194 +0000 UTC m=+1224.956404018" watchObservedRunningTime="2026-01-03 04:01:29.345886872 +0000 UTC m=+1224.957313696" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.313360 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.316091 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.326998 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.327068 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.352447 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.362694 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.377681 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.378561 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.400052 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:38 crc kubenswrapper[4921]: I0103 04:01:38.400914 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:39 crc kubenswrapper[4921]: I0103 04:01:39.379898 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:39 crc kubenswrapper[4921]: I0103 04:01:39.380379 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:40 crc kubenswrapper[4921]: I0103 04:01:40.387494 4921 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 03 04:01:40 crc kubenswrapper[4921]: I0103 04:01:40.387833 4921 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 03 04:01:40 crc kubenswrapper[4921]: I0103 04:01:40.721777 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:40 crc kubenswrapper[4921]: I0103 04:01:40.730763 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:40 crc kubenswrapper[4921]: I0103 04:01:40.732450 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:41 crc kubenswrapper[4921]: I0103 04:01:41.363117 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:01:41 crc kubenswrapper[4921]: I0103 04:01:41.418850 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:43 crc kubenswrapper[4921]: I0103 04:01:43.410261 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-log" containerID="cri-o://f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e" gracePeriod=30 Jan 03 04:01:43 crc kubenswrapper[4921]: I0103 04:01:43.410397 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-httpd" containerID="cri-o://c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc" gracePeriod=30 Jan 03 04:01:45 crc kubenswrapper[4921]: I0103 04:01:45.425412 4921 generic.go:334] "Generic (PLEG): container finished" podID="728242d3-f998-4c07-af63-565958c51a56" containerID="f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e" exitCode=143 Jan 03 04:01:45 crc kubenswrapper[4921]: I0103 04:01:45.425677 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerDied","Data":"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e"} Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.036259 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190419 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190460 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190493 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190516 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190554 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190593 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190611 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190625 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190642 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190667 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqpt8\" (UniqueName: \"kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190685 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190698 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190743 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.190758 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick\") pod \"728242d3-f998-4c07-af63-565958c51a56\" (UID: \"728242d3-f998-4c07-af63-565958c51a56\") " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191041 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191346 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191556 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs" (OuterVolumeSpecName: "logs") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191588 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys" (OuterVolumeSpecName: "sys") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191607 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run" (OuterVolumeSpecName: "run") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.191622 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev" (OuterVolumeSpecName: "dev") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.192318 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.192380 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.192378 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.196983 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8" (OuterVolumeSpecName: "kube-api-access-fqpt8") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "kube-api-access-fqpt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.197613 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.199444 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance-cache") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.199998 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts" (OuterVolumeSpecName: "scripts") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.252722 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data" (OuterVolumeSpecName: "config-data") pod "728242d3-f998-4c07-af63-565958c51a56" (UID: "728242d3-f998-4c07-af63-565958c51a56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292530 4921 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292575 4921 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/728242d3-f998-4c07-af63-565958c51a56-logs\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292587 4921 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-sys\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292598 4921 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292610 4921 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-dev\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292620 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292634 4921 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292645 4921 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292656 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/728242d3-f998-4c07-af63-565958c51a56-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292671 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqpt8\" (UniqueName: \"kubernetes.io/projected/728242d3-f998-4c07-af63-565958c51a56-kube-api-access-fqpt8\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292712 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292725 4921 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292742 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.292753 4921 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/728242d3-f998-4c07-af63-565958c51a56-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.306880 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.312472 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.394481 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.394516 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.448963 4921 generic.go:334] "Generic (PLEG): container finished" podID="728242d3-f998-4c07-af63-565958c51a56" containerID="c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc" exitCode=0 Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.449003 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerDied","Data":"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc"} Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.449035 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"728242d3-f998-4c07-af63-565958c51a56","Type":"ContainerDied","Data":"d33b2e77eb686304f785f2fae83eb1f1e05041fa247ee11ee8e53ed49a6fa1ce"} Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.449053 4921 scope.go:117] "RemoveContainer" containerID="c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.449066 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.489569 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.497606 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.500249 4921 scope.go:117] "RemoveContainer" containerID="f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.520699 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:47 crc kubenswrapper[4921]: E0103 04:01:47.521007 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-log" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.521024 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-log" Jan 03 04:01:47 crc kubenswrapper[4921]: E0103 04:01:47.521037 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-httpd" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.521042 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-httpd" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.521176 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-httpd" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.521189 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="728242d3-f998-4c07-af63-565958c51a56" containerName="glance-log" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.522051 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.536041 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.543465 4921 scope.go:117] "RemoveContainer" containerID="c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc" Jan 03 04:01:47 crc kubenswrapper[4921]: E0103 04:01:47.544145 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc\": container with ID starting with c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc not found: ID does not exist" containerID="c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.544198 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc"} err="failed to get container status \"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc\": rpc error: code = NotFound desc = could not find container \"c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc\": container with ID starting with c95f88063da2004b375593fa72a96ad06c298f2cfee83b609171ba03b2da20dc not found: ID does not exist" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.544235 4921 scope.go:117] "RemoveContainer" containerID="f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e" Jan 03 04:01:47 crc kubenswrapper[4921]: E0103 04:01:47.544675 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e\": container with ID starting with f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e not found: ID does not exist" containerID="f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.544722 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e"} err="failed to get container status \"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e\": rpc error: code = NotFound desc = could not find container \"f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e\": container with ID starting with f927dc7879026e9721bc47ee8b48c9bc7e0ea9690936d523453d609f637b259e not found: ID does not exist" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699034 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699076 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699097 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699131 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699152 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699258 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699292 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699308 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699322 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699345 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699359 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699392 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699441 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rcqz\" (UniqueName: \"kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.699492 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800478 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800523 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800559 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800586 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rcqz\" (UniqueName: \"kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800616 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800633 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800650 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800667 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800695 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800714 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800734 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800754 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800771 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800785 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800848 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800881 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800902 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.800985 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.803744 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.803864 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.803909 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.804201 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.804438 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.804936 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.805038 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.814628 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.814846 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.846014 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rcqz\" (UniqueName: \"kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.851512 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:47 crc kubenswrapper[4921]: I0103 04:01:47.855464 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-single-0\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:48 crc kubenswrapper[4921]: I0103 04:01:48.150085 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:48 crc kubenswrapper[4921]: I0103 04:01:48.607590 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:01:48 crc kubenswrapper[4921]: W0103 04:01:48.619671 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod742050ad_7ef5_4e43_8267_8e8d91d5f140.slice/crio-f6b9cbb60d5c30e81408e3aedd0797a517b520497e2b231306c320197a81b4d3 WatchSource:0}: Error finding container f6b9cbb60d5c30e81408e3aedd0797a517b520497e2b231306c320197a81b4d3: Status 404 returned error can't find the container with id f6b9cbb60d5c30e81408e3aedd0797a517b520497e2b231306c320197a81b4d3 Jan 03 04:01:48 crc kubenswrapper[4921]: I0103 04:01:48.901475 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="728242d3-f998-4c07-af63-565958c51a56" path="/var/lib/kubelet/pods/728242d3-f998-4c07-af63-565958c51a56/volumes" Jan 03 04:01:49 crc kubenswrapper[4921]: I0103 04:01:49.469324 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerStarted","Data":"36bcbb7bd15dbefdd8ecea082f2f3e2e7b4a3426b4b963d6e944861c5ea628ae"} Jan 03 04:01:49 crc kubenswrapper[4921]: I0103 04:01:49.469979 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerStarted","Data":"d88924594f52575e03ae17a918f011af35601176ada1f4e5fa3651e912e040c8"} Jan 03 04:01:49 crc kubenswrapper[4921]: I0103 04:01:49.469997 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerStarted","Data":"f6b9cbb60d5c30e81408e3aedd0797a517b520497e2b231306c320197a81b4d3"} Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.150945 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.151717 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.193701 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.210635 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.229126 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=11.229100019 podStartE2EDuration="11.229100019s" podCreationTimestamp="2026-01-03 04:01:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:01:49.51487805 +0000 UTC m=+1245.126304924" watchObservedRunningTime="2026-01-03 04:01:58.229100019 +0000 UTC m=+1253.840526853" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.543884 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:01:58 crc kubenswrapper[4921]: I0103 04:01:58.543944 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:00 crc kubenswrapper[4921]: I0103 04:02:00.523085 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:00 crc kubenswrapper[4921]: I0103 04:02:00.523532 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.143367 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6c824"] Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.151888 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-6c824"] Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.260379 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-dw4rz"] Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.264689 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.271410 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.271554 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.278573 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-dw4rz"] Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.367060 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.367311 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4xcd\" (UniqueName: \"kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.367424 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.367497 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.468450 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4xcd\" (UniqueName: \"kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.468511 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.468534 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.468581 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.473895 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.474413 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.478632 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.488916 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4xcd\" (UniqueName: \"kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd\") pod \"glance-db-sync-dw4rz\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:13 crc kubenswrapper[4921]: I0103 04:02:13.602686 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:14 crc kubenswrapper[4921]: I0103 04:02:14.094733 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-dw4rz"] Jan 03 04:02:14 crc kubenswrapper[4921]: W0103 04:02:14.101528 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fe244be_5048_438b_8686_471cb8d5cc2e.slice/crio-4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64 WatchSource:0}: Error finding container 4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64: Status 404 returned error can't find the container with id 4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64 Jan 03 04:02:14 crc kubenswrapper[4921]: I0103 04:02:14.755041 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-dw4rz" event={"ID":"3fe244be-5048-438b-8686-471cb8d5cc2e","Type":"ContainerStarted","Data":"9b44bb323f503fc1bfcf41778be5c4f7739c8569759db79bbfa72e0003f514c4"} Jan 03 04:02:14 crc kubenswrapper[4921]: I0103 04:02:14.755458 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-dw4rz" event={"ID":"3fe244be-5048-438b-8686-471cb8d5cc2e","Type":"ContainerStarted","Data":"4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64"} Jan 03 04:02:14 crc kubenswrapper[4921]: I0103 04:02:14.775025 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-dw4rz" podStartSLOduration=1.775002199 podStartE2EDuration="1.775002199s" podCreationTimestamp="2026-01-03 04:02:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:02:14.768867563 +0000 UTC m=+1270.380294427" watchObservedRunningTime="2026-01-03 04:02:14.775002199 +0000 UTC m=+1270.386429023" Jan 03 04:02:14 crc kubenswrapper[4921]: I0103 04:02:14.893020 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57fda7f2-4f05-4abb-80c3-686dda011b05" path="/var/lib/kubelet/pods/57fda7f2-4f05-4abb-80c3-686dda011b05/volumes" Jan 03 04:02:17 crc kubenswrapper[4921]: I0103 04:02:17.782954 4921 generic.go:334] "Generic (PLEG): container finished" podID="3fe244be-5048-438b-8686-471cb8d5cc2e" containerID="9b44bb323f503fc1bfcf41778be5c4f7739c8569759db79bbfa72e0003f514c4" exitCode=0 Jan 03 04:02:17 crc kubenswrapper[4921]: I0103 04:02:17.783046 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-dw4rz" event={"ID":"3fe244be-5048-438b-8686-471cb8d5cc2e","Type":"ContainerDied","Data":"9b44bb323f503fc1bfcf41778be5c4f7739c8569759db79bbfa72e0003f514c4"} Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.167895 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.362113 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4xcd\" (UniqueName: \"kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd\") pod \"3fe244be-5048-438b-8686-471cb8d5cc2e\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.362526 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle\") pod \"3fe244be-5048-438b-8686-471cb8d5cc2e\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.362579 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data\") pod \"3fe244be-5048-438b-8686-471cb8d5cc2e\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.362614 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data\") pod \"3fe244be-5048-438b-8686-471cb8d5cc2e\" (UID: \"3fe244be-5048-438b-8686-471cb8d5cc2e\") " Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.378220 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3fe244be-5048-438b-8686-471cb8d5cc2e" (UID: "3fe244be-5048-438b-8686-471cb8d5cc2e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.379213 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd" (OuterVolumeSpecName: "kube-api-access-n4xcd") pod "3fe244be-5048-438b-8686-471cb8d5cc2e" (UID: "3fe244be-5048-438b-8686-471cb8d5cc2e"). InnerVolumeSpecName "kube-api-access-n4xcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.400065 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fe244be-5048-438b-8686-471cb8d5cc2e" (UID: "3fe244be-5048-438b-8686-471cb8d5cc2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.428013 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data" (OuterVolumeSpecName: "config-data") pod "3fe244be-5048-438b-8686-471cb8d5cc2e" (UID: "3fe244be-5048-438b-8686-471cb8d5cc2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.464840 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4xcd\" (UniqueName: \"kubernetes.io/projected/3fe244be-5048-438b-8686-471cb8d5cc2e-kube-api-access-n4xcd\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.464886 4921 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.464905 4921 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.464922 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fe244be-5048-438b-8686-471cb8d5cc2e-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.801828 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-dw4rz" event={"ID":"3fe244be-5048-438b-8686-471cb8d5cc2e","Type":"ContainerDied","Data":"4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64"} Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.801859 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-dw4rz" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.801907 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4521960d726571fbf0cf2ad6f41a3850b7b7f548a6496a88cc00cc7431b35a64" Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.991100 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.991480 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-log" containerID="cri-o://83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c" gracePeriod=30 Jan 03 04:02:19 crc kubenswrapper[4921]: I0103 04:02:19.992028 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-httpd" containerID="cri-o://86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58" gracePeriod=30 Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.052010 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.052620 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-log" containerID="cri-o://d88924594f52575e03ae17a918f011af35601176ada1f4e5fa3651e912e040c8" gracePeriod=30 Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.053359 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-httpd" containerID="cri-o://36bcbb7bd15dbefdd8ecea082f2f3e2e7b4a3426b4b963d6e944861c5ea628ae" gracePeriod=30 Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.069490 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.105:9292/healthcheck\": EOF" Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.069542 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.105:9292/healthcheck\": EOF" Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.813140 4921 generic.go:334] "Generic (PLEG): container finished" podID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerID="83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c" exitCode=143 Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.813296 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerDied","Data":"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c"} Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.816225 4921 generic.go:334] "Generic (PLEG): container finished" podID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerID="d88924594f52575e03ae17a918f011af35601176ada1f4e5fa3651e912e040c8" exitCode=143 Jan 03 04:02:20 crc kubenswrapper[4921]: I0103 04:02:20.816325 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerDied","Data":"d88924594f52575e03ae17a918f011af35601176ada1f4e5fa3651e912e040c8"} Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.643058 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829004 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829059 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829082 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829119 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829147 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829204 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829222 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829250 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtq57\" (UniqueName: \"kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829264 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829300 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829325 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829372 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829529 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829557 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev" (OuterVolumeSpecName: "dev") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829520 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829578 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.829590 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys" (OuterVolumeSpecName: "sys") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830190 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs" (OuterVolumeSpecName: "logs") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830237 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830255 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run\") pod \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\" (UID: \"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4\") " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830325 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830543 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830581 4921 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830597 4921 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-logs\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830606 4921 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830615 4921 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-dev\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830623 4921 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-sys\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830632 4921 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830644 4921 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.830680 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run" (OuterVolumeSpecName: "run") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.835461 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts" (OuterVolumeSpecName: "scripts") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.836403 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57" (OuterVolumeSpecName: "kube-api-access-qtq57") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "kube-api-access-qtq57". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.837087 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.845758 4921 generic.go:334] "Generic (PLEG): container finished" podID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerID="86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58" exitCode=0 Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.845815 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.845832 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerDied","Data":"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58"} Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.845859 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4","Type":"ContainerDied","Data":"ee5fe17bad747878c86f86cd362f85f4b3f62e11f106bdfee3a869de56513618"} Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.845883 4921 scope.go:117] "RemoveContainer" containerID="86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.849452 4921 generic.go:334] "Generic (PLEG): container finished" podID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerID="36bcbb7bd15dbefdd8ecea082f2f3e2e7b4a3426b4b963d6e944861c5ea628ae" exitCode=0 Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.849488 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerDied","Data":"36bcbb7bd15dbefdd8ecea082f2f3e2e7b4a3426b4b963d6e944861c5ea628ae"} Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.849895 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "glance") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.865665 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data" (OuterVolumeSpecName: "config-data") pod "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" (UID: "f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.891992 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.900167 4921 scope.go:117] "RemoveContainer" containerID="83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.923217 4921 scope.go:117] "RemoveContainer" containerID="86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58" Jan 03 04:02:23 crc kubenswrapper[4921]: E0103 04:02:23.923593 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58\": container with ID starting with 86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58 not found: ID does not exist" containerID="86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.923630 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58"} err="failed to get container status \"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58\": rpc error: code = NotFound desc = could not find container \"86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58\": container with ID starting with 86f79362891586f66fc99975d1c719938aa8da076521285bcdbc517e52626e58 not found: ID does not exist" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.923655 4921 scope.go:117] "RemoveContainer" containerID="83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c" Jan 03 04:02:23 crc kubenswrapper[4921]: E0103 04:02:23.923954 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c\": container with ID starting with 83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c not found: ID does not exist" containerID="83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.923998 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c"} err="failed to get container status \"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c\": rpc error: code = NotFound desc = could not find container \"83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c\": container with ID starting with 83ab82867ff2bf51a1748c543ac3cfad80fcfd2ccc0922e83412b965899c047c not found: ID does not exist" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.931968 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.931993 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtq57\" (UniqueName: \"kubernetes.io/projected/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-kube-api-access-qtq57\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.932023 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.932045 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.932055 4921 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.932064 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.932074 4921 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.950722 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Jan 03 04:02:23 crc kubenswrapper[4921]: I0103 04:02:23.959999 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032563 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032617 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032657 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032674 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032696 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev" (OuterVolumeSpecName: "dev") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032716 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032746 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032767 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys" (OuterVolumeSpecName: "sys") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032779 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032851 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rcqz\" (UniqueName: \"kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032792 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032876 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032826 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run" (OuterVolumeSpecName: "run") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032915 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032936 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032935 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032977 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.032998 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033004 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033066 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033129 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data\") pod \"742050ad-7ef5-4e43-8267-8e8d91d5f140\" (UID: \"742050ad-7ef5-4e43-8267-8e8d91d5f140\") " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033353 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033776 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs" (OuterVolumeSpecName: "logs") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033793 4921 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-dev\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033816 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033829 4921 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-sys\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033840 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033851 4921 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033862 4921 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033872 4921 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033883 4921 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033893 4921 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/742050ad-7ef5-4e43-8267-8e8d91d5f140-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.033904 4921 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.035156 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.035834 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance-cache") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.035943 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts" (OuterVolumeSpecName: "scripts") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.036433 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz" (OuterVolumeSpecName: "kube-api-access-9rcqz") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "kube-api-access-9rcqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.082030 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data" (OuterVolumeSpecName: "config-data") pod "742050ad-7ef5-4e43-8267-8e8d91d5f140" (UID: "742050ad-7ef5-4e43-8267-8e8d91d5f140"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.135621 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.135678 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.136184 4921 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/742050ad-7ef5-4e43-8267-8e8d91d5f140-logs\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.136240 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rcqz\" (UniqueName: \"kubernetes.io/projected/742050ad-7ef5-4e43-8267-8e8d91d5f140-kube-api-access-9rcqz\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.136321 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.136348 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/742050ad-7ef5-4e43-8267-8e8d91d5f140-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.154626 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.166803 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.179506 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.185159 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.238235 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.238302 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.865482 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"742050ad-7ef5-4e43-8267-8e8d91d5f140","Type":"ContainerDied","Data":"f6b9cbb60d5c30e81408e3aedd0797a517b520497e2b231306c320197a81b4d3"} Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.865560 4921 scope.go:117] "RemoveContainer" containerID="36bcbb7bd15dbefdd8ecea082f2f3e2e7b4a3426b4b963d6e944861c5ea628ae" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.865785 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.900789 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" path="/var/lib/kubelet/pods/f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4/volumes" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.906783 4921 scope.go:117] "RemoveContainer" containerID="d88924594f52575e03ae17a918f011af35601176ada1f4e5fa3651e912e040c8" Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.918098 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:24 crc kubenswrapper[4921]: I0103 04:02:24.926992 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.399329 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:25 crc kubenswrapper[4921]: E0103 04:02:25.399871 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.399910 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: E0103 04:02:25.399940 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.399956 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: E0103 04:02:25.399980 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400000 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: E0103 04:02:25.400022 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe244be-5048-438b-8686-471cb8d5cc2e" containerName="glance-db-sync" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400039 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe244be-5048-438b-8686-471cb8d5cc2e" containerName="glance-db-sync" Jan 03 04:02:25 crc kubenswrapper[4921]: E0103 04:02:25.400077 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400092 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400450 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400480 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-log" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400517 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400547 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe244be-5048-438b-8686-471cb8d5cc2e" containerName="glance-db-sync" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.400573 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0fcf55e-ea3e-4b3a-b06f-dfd649e8c0e4" containerName="glance-httpd" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.402610 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.407030 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.407776 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.410329 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.410929 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.410960 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.417873 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-ddhvq" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.432159 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556503 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556540 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzsrg\" (UniqueName: \"kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556566 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556599 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556682 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556718 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556738 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556756 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.556774 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.657872 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.657915 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzsrg\" (UniqueName: \"kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.657944 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.657978 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658011 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658050 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658080 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658106 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658129 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658740 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658801 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.658951 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.664260 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.664977 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.670012 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.675468 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.676619 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.680604 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzsrg\" (UniqueName: \"kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.692658 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-single-0\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:25 crc kubenswrapper[4921]: I0103 04:02:25.729229 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:26 crc kubenswrapper[4921]: I0103 04:02:26.162910 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:26 crc kubenswrapper[4921]: I0103 04:02:26.923008 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742050ad-7ef5-4e43-8267-8e8d91d5f140" path="/var/lib/kubelet/pods/742050ad-7ef5-4e43-8267-8e8d91d5f140/volumes" Jan 03 04:02:26 crc kubenswrapper[4921]: I0103 04:02:26.939443 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerStarted","Data":"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc"} Jan 03 04:02:26 crc kubenswrapper[4921]: I0103 04:02:26.939491 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerStarted","Data":"5d33f179b5eed2fd0b0ebbeb555fe5849db5f01ef70d4f78a435edbab91469c5"} Jan 03 04:02:27 crc kubenswrapper[4921]: I0103 04:02:27.950455 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerStarted","Data":"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730"} Jan 03 04:02:27 crc kubenswrapper[4921]: I0103 04:02:27.977944 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.977921704 podStartE2EDuration="2.977921704s" podCreationTimestamp="2026-01-03 04:02:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:02:27.977691978 +0000 UTC m=+1283.589118822" watchObservedRunningTime="2026-01-03 04:02:27.977921704 +0000 UTC m=+1283.589348538" Jan 03 04:02:35 crc kubenswrapper[4921]: I0103 04:02:35.729522 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:35 crc kubenswrapper[4921]: I0103 04:02:35.729959 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:35 crc kubenswrapper[4921]: I0103 04:02:35.775940 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:35 crc kubenswrapper[4921]: I0103 04:02:35.785703 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:36 crc kubenswrapper[4921]: I0103 04:02:36.034442 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:36 crc kubenswrapper[4921]: I0103 04:02:36.034496 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:37 crc kubenswrapper[4921]: I0103 04:02:37.825764 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:37 crc kubenswrapper[4921]: I0103 04:02:37.930130 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.835125 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-dw4rz"] Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.842502 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-dw4rz"] Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.920660 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe244be-5048-438b-8686-471cb8d5cc2e" path="/var/lib/kubelet/pods/3fe244be-5048-438b-8686-471cb8d5cc2e/volumes" Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.921746 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancea6c3-account-delete-p2vfn"] Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.922927 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.925385 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea6c3-account-delete-p2vfn"] Jan 03 04:02:38 crc kubenswrapper[4921]: I0103 04:02:38.933779 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.016610 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jphnf\" (UniqueName: \"kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf\") pod \"glancea6c3-account-delete-p2vfn\" (UID: \"4a616f54-4b6a-4108-ad6f-6fb822a6f12a\") " pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.076881 4921 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="glance-kuttl-tests/glance-default-single-0" secret="" err="secret \"glance-glance-dockercfg-ddhvq\" not found" Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.118672 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jphnf\" (UniqueName: \"kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf\") pod \"glancea6c3-account-delete-p2vfn\" (UID: \"4a616f54-4b6a-4108-ad6f-6fb822a6f12a\") " pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.118697 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.118758 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.118762 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:39.618742509 +0000 UTC m=+1295.230169333 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-default-single-config-data" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.118899 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:39.618882333 +0000 UTC m=+1295.230309157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-scripts" not found Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.154823 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jphnf\" (UniqueName: \"kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf\") pod \"glancea6c3-account-delete-p2vfn\" (UID: \"4a616f54-4b6a-4108-ad6f-6fb822a6f12a\") " pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.249679 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:39 crc kubenswrapper[4921]: I0103 04:02:39.533883 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancea6c3-account-delete-p2vfn"] Jan 03 04:02:39 crc kubenswrapper[4921]: W0103 04:02:39.536540 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a616f54_4b6a_4108_ad6f_6fb822a6f12a.slice/crio-e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a WatchSource:0}: Error finding container e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a: Status 404 returned error can't find the container with id e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.626713 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.626772 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:40.626759461 +0000 UTC m=+1296.238186285 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-scripts" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.626815 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Jan 03 04:02:39 crc kubenswrapper[4921]: E0103 04:02:39.626835 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:40.626828423 +0000 UTC m=+1296.238255247 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-default-single-config-data" not found Jan 03 04:02:40 crc kubenswrapper[4921]: I0103 04:02:40.092014 4921 generic.go:334] "Generic (PLEG): container finished" podID="4a616f54-4b6a-4108-ad6f-6fb822a6f12a" containerID="b0476718bfde774e30f4630e50fbac430b23e66798d51049f7989a1800f73e81" exitCode=0 Jan 03 04:02:40 crc kubenswrapper[4921]: I0103 04:02:40.092414 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" event={"ID":"4a616f54-4b6a-4108-ad6f-6fb822a6f12a","Type":"ContainerDied","Data":"b0476718bfde774e30f4630e50fbac430b23e66798d51049f7989a1800f73e81"} Jan 03 04:02:40 crc kubenswrapper[4921]: I0103 04:02:40.092488 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" event={"ID":"4a616f54-4b6a-4108-ad6f-6fb822a6f12a","Type":"ContainerStarted","Data":"e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a"} Jan 03 04:02:40 crc kubenswrapper[4921]: I0103 04:02:40.092684 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-log" containerID="cri-o://bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc" gracePeriod=30 Jan 03 04:02:40 crc kubenswrapper[4921]: I0103 04:02:40.093333 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-httpd" containerID="cri-o://1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730" gracePeriod=30 Jan 03 04:02:40 crc kubenswrapper[4921]: E0103 04:02:40.639801 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Jan 03 04:02:40 crc kubenswrapper[4921]: E0103 04:02:40.639884 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:42.63986799 +0000 UTC m=+1298.251294814 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-scripts" not found Jan 03 04:02:40 crc kubenswrapper[4921]: E0103 04:02:40.639894 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Jan 03 04:02:40 crc kubenswrapper[4921]: E0103 04:02:40.640040 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:42.640000194 +0000 UTC m=+1298.251427048 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-default-single-config-data" not found Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.106577 4921 generic.go:334] "Generic (PLEG): container finished" podID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerID="bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc" exitCode=143 Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.106655 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerDied","Data":"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc"} Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.442719 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.453822 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jphnf\" (UniqueName: \"kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf\") pod \"4a616f54-4b6a-4108-ad6f-6fb822a6f12a\" (UID: \"4a616f54-4b6a-4108-ad6f-6fb822a6f12a\") " Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.464136 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf" (OuterVolumeSpecName: "kube-api-access-jphnf") pod "4a616f54-4b6a-4108-ad6f-6fb822a6f12a" (UID: "4a616f54-4b6a-4108-ad6f-6fb822a6f12a"). InnerVolumeSpecName "kube-api-access-jphnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:41 crc kubenswrapper[4921]: I0103 04:02:41.555722 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jphnf\" (UniqueName: \"kubernetes.io/projected/4a616f54-4b6a-4108-ad6f-6fb822a6f12a-kube-api-access-jphnf\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:42 crc kubenswrapper[4921]: I0103 04:02:42.117919 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" event={"ID":"4a616f54-4b6a-4108-ad6f-6fb822a6f12a","Type":"ContainerDied","Data":"e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a"} Jan 03 04:02:42 crc kubenswrapper[4921]: I0103 04:02:42.117973 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancea6c3-account-delete-p2vfn" Jan 03 04:02:42 crc kubenswrapper[4921]: I0103 04:02:42.117986 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e96507254a23c1a35489f98fd46976d8a4babd2053ba89010f3f1cc5c9e4f27a" Jan 03 04:02:42 crc kubenswrapper[4921]: E0103 04:02:42.672001 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Jan 03 04:02:42 crc kubenswrapper[4921]: E0103 04:02:42.672314 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:46.672296279 +0000 UTC m=+1302.283723113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-scripts" not found Jan 03 04:02:42 crc kubenswrapper[4921]: E0103 04:02:42.672010 4921 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Jan 03 04:02:42 crc kubenswrapper[4921]: E0103 04:02:42.672412 4921 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data podName:7309e473-5c23-4bba-a6ac-c2f18e3d0a2e nodeName:}" failed. No retries permitted until 2026-01-03 04:02:46.672384681 +0000 UTC m=+1302.283811545 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data") pod "glance-default-single-0" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e") : secret "glance-default-single-config-data" not found Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.692038 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789093 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789432 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789482 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789505 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789524 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzsrg\" (UniqueName: \"kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789556 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789571 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789606 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789620 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\" (UID: \"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e\") " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.789559 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.790439 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs" (OuterVolumeSpecName: "logs") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.799706 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts" (OuterVolumeSpecName: "scripts") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.800327 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg" (OuterVolumeSpecName: "kube-api-access-pzsrg") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "kube-api-access-pzsrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.806357 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.813536 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.823411 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.825066 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.841411 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data" (OuterVolumeSpecName: "config-data") pod "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" (UID: "7309e473-5c23-4bba-a6ac-c2f18e3d0a2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891217 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891253 4921 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891282 4921 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-logs\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891317 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891327 4921 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891335 4921 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891343 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891351 4921 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.891359 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzsrg\" (UniqueName: \"kubernetes.io/projected/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e-kube-api-access-pzsrg\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.895494 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-b47hv"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.904351 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-b47hv"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.910666 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.913774 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancea6c3-account-delete-p2vfn"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.920014 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-a6c3-account-create-9cb66"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.924576 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancea6c3-account-delete-p2vfn"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.929021 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-a6c3-account-create-9cb66"] Jan 03 04:02:43 crc kubenswrapper[4921]: I0103 04:02:43.993071 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.139239 4921 generic.go:334] "Generic (PLEG): container finished" podID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerID="1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730" exitCode=0 Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.139299 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerDied","Data":"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730"} Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.139345 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7309e473-5c23-4bba-a6ac-c2f18e3d0a2e","Type":"ContainerDied","Data":"5d33f179b5eed2fd0b0ebbeb555fe5849db5f01ef70d4f78a435edbab91469c5"} Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.139365 4921 scope.go:117] "RemoveContainer" containerID="1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.139374 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.172512 4921 scope.go:117] "RemoveContainer" containerID="bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.179257 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.186660 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.194510 4921 scope.go:117] "RemoveContainer" containerID="1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730" Jan 03 04:02:44 crc kubenswrapper[4921]: E0103 04:02:44.194993 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730\": container with ID starting with 1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730 not found: ID does not exist" containerID="1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.195037 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730"} err="failed to get container status \"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730\": rpc error: code = NotFound desc = could not find container \"1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730\": container with ID starting with 1d22aad4dcee24d15ebf05ea1689f9e2b2394cdf43a2fa8c073dcc7fd6a87730 not found: ID does not exist" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.195071 4921 scope.go:117] "RemoveContainer" containerID="bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc" Jan 03 04:02:44 crc kubenswrapper[4921]: E0103 04:02:44.195346 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc\": container with ID starting with bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc not found: ID does not exist" containerID="bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.195404 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc"} err="failed to get container status \"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc\": rpc error: code = NotFound desc = could not find container \"bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc\": container with ID starting with bbc9e33600b8c3f68924ac6d15c9e1c49668efe30f524c0859735c94b8fb61bc not found: ID does not exist" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.897576 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0409c8d5-d153-4227-99c2-45414648ba2c" path="/var/lib/kubelet/pods/0409c8d5-d153-4227-99c2-45414648ba2c/volumes" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.898474 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a616f54-4b6a-4108-ad6f-6fb822a6f12a" path="/var/lib/kubelet/pods/4a616f54-4b6a-4108-ad6f-6fb822a6f12a/volumes" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.899068 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" path="/var/lib/kubelet/pods/7309e473-5c23-4bba-a6ac-c2f18e3d0a2e/volumes" Jan 03 04:02:44 crc kubenswrapper[4921]: I0103 04:02:44.900166 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf030d9-6b0b-4992-b355-68c358153bce" path="/var/lib/kubelet/pods/daf030d9-6b0b-4992-b355-68c358153bce/volumes" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.730799 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-wnxlc"] Jan 03 04:02:45 crc kubenswrapper[4921]: E0103 04:02:45.731206 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a616f54-4b6a-4108-ad6f-6fb822a6f12a" containerName="mariadb-account-delete" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731231 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a616f54-4b6a-4108-ad6f-6fb822a6f12a" containerName="mariadb-account-delete" Jan 03 04:02:45 crc kubenswrapper[4921]: E0103 04:02:45.731249 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-httpd" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731258 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-httpd" Jan 03 04:02:45 crc kubenswrapper[4921]: E0103 04:02:45.731301 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-log" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731310 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-log" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731459 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a616f54-4b6a-4108-ad6f-6fb822a6f12a" containerName="mariadb-account-delete" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731472 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-httpd" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.731498 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="7309e473-5c23-4bba-a6ac-c2f18e3d0a2e" containerName="glance-log" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.732000 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.738803 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-wnxlc"] Jan 03 04:02:45 crc kubenswrapper[4921]: I0103 04:02:45.919621 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdl6s\" (UniqueName: \"kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s\") pod \"glance-db-create-wnxlc\" (UID: \"b1054977-0f34-413d-958f-5cdfe6c881f7\") " pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:46 crc kubenswrapper[4921]: I0103 04:02:46.021787 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdl6s\" (UniqueName: \"kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s\") pod \"glance-db-create-wnxlc\" (UID: \"b1054977-0f34-413d-958f-5cdfe6c881f7\") " pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:46 crc kubenswrapper[4921]: I0103 04:02:46.056546 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdl6s\" (UniqueName: \"kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s\") pod \"glance-db-create-wnxlc\" (UID: \"b1054977-0f34-413d-958f-5cdfe6c881f7\") " pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:46 crc kubenswrapper[4921]: I0103 04:02:46.062069 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:46 crc kubenswrapper[4921]: I0103 04:02:46.605024 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-wnxlc"] Jan 03 04:02:47 crc kubenswrapper[4921]: I0103 04:02:47.165807 4921 generic.go:334] "Generic (PLEG): container finished" podID="b1054977-0f34-413d-958f-5cdfe6c881f7" containerID="49202dc7df1ec6ce9150f547f9fc05bd234450bb71407f0429d6183ccc65412e" exitCode=0 Jan 03 04:02:47 crc kubenswrapper[4921]: I0103 04:02:47.165860 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-wnxlc" event={"ID":"b1054977-0f34-413d-958f-5cdfe6c881f7","Type":"ContainerDied","Data":"49202dc7df1ec6ce9150f547f9fc05bd234450bb71407f0429d6183ccc65412e"} Jan 03 04:02:47 crc kubenswrapper[4921]: I0103 04:02:47.166115 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-wnxlc" event={"ID":"b1054977-0f34-413d-958f-5cdfe6c881f7","Type":"ContainerStarted","Data":"f4cfc8d3c36b1a70685d7c52597ffeb5e43928f795d1378d910e86924f6cce46"} Jan 03 04:02:48 crc kubenswrapper[4921]: I0103 04:02:48.528213 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:48 crc kubenswrapper[4921]: I0103 04:02:48.662748 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdl6s\" (UniqueName: \"kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s\") pod \"b1054977-0f34-413d-958f-5cdfe6c881f7\" (UID: \"b1054977-0f34-413d-958f-5cdfe6c881f7\") " Jan 03 04:02:48 crc kubenswrapper[4921]: I0103 04:02:48.667983 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s" (OuterVolumeSpecName: "kube-api-access-tdl6s") pod "b1054977-0f34-413d-958f-5cdfe6c881f7" (UID: "b1054977-0f34-413d-958f-5cdfe6c881f7"). InnerVolumeSpecName "kube-api-access-tdl6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:48 crc kubenswrapper[4921]: I0103 04:02:48.764935 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdl6s\" (UniqueName: \"kubernetes.io/projected/b1054977-0f34-413d-958f-5cdfe6c881f7-kube-api-access-tdl6s\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:49 crc kubenswrapper[4921]: I0103 04:02:49.187908 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-wnxlc" event={"ID":"b1054977-0f34-413d-958f-5cdfe6c881f7","Type":"ContainerDied","Data":"f4cfc8d3c36b1a70685d7c52597ffeb5e43928f795d1378d910e86924f6cce46"} Jan 03 04:02:49 crc kubenswrapper[4921]: I0103 04:02:49.188320 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4cfc8d3c36b1a70685d7c52597ffeb5e43928f795d1378d910e86924f6cce46" Jan 03 04:02:49 crc kubenswrapper[4921]: I0103 04:02:49.187964 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-wnxlc" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.748006 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-09aa-account-create-qmd7n"] Jan 03 04:02:55 crc kubenswrapper[4921]: E0103 04:02:55.748687 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1054977-0f34-413d-958f-5cdfe6c881f7" containerName="mariadb-database-create" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.748700 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1054977-0f34-413d-958f-5cdfe6c881f7" containerName="mariadb-database-create" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.748825 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1054977-0f34-413d-958f-5cdfe6c881f7" containerName="mariadb-database-create" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.749295 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.751339 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.760038 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-09aa-account-create-qmd7n"] Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.886374 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzc9s\" (UniqueName: \"kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s\") pod \"glance-09aa-account-create-qmd7n\" (UID: \"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00\") " pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:55 crc kubenswrapper[4921]: I0103 04:02:55.987948 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzc9s\" (UniqueName: \"kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s\") pod \"glance-09aa-account-create-qmd7n\" (UID: \"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00\") " pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:56 crc kubenswrapper[4921]: I0103 04:02:56.011820 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzc9s\" (UniqueName: \"kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s\") pod \"glance-09aa-account-create-qmd7n\" (UID: \"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00\") " pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:56 crc kubenswrapper[4921]: I0103 04:02:56.072227 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:56 crc kubenswrapper[4921]: I0103 04:02:56.599612 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-09aa-account-create-qmd7n"] Jan 03 04:02:57 crc kubenswrapper[4921]: I0103 04:02:57.285770 4921 generic.go:334] "Generic (PLEG): container finished" podID="015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" containerID="3553b26bf3b71df953c4ffe305d0272e0c5ff024c72e315af8b2a89535517cc2" exitCode=0 Jan 03 04:02:57 crc kubenswrapper[4921]: I0103 04:02:57.285922 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" event={"ID":"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00","Type":"ContainerDied","Data":"3553b26bf3b71df953c4ffe305d0272e0c5ff024c72e315af8b2a89535517cc2"} Jan 03 04:02:57 crc kubenswrapper[4921]: I0103 04:02:57.286052 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" event={"ID":"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00","Type":"ContainerStarted","Data":"0592a8755f6a106db2fb1c7ba296474e381859b65e2814d453a87a5362c50e9b"} Jan 03 04:02:58 crc kubenswrapper[4921]: I0103 04:02:58.641338 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:02:58 crc kubenswrapper[4921]: I0103 04:02:58.838346 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzc9s\" (UniqueName: \"kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s\") pod \"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00\" (UID: \"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00\") " Jan 03 04:02:58 crc kubenswrapper[4921]: I0103 04:02:58.848937 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s" (OuterVolumeSpecName: "kube-api-access-qzc9s") pod "015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" (UID: "015d13a2-fc39-4b75-b4ec-8d5bf46c4c00"). InnerVolumeSpecName "kube-api-access-qzc9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:02:58 crc kubenswrapper[4921]: I0103 04:02:58.939924 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzc9s\" (UniqueName: \"kubernetes.io/projected/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00-kube-api-access-qzc9s\") on node \"crc\" DevicePath \"\"" Jan 03 04:02:59 crc kubenswrapper[4921]: I0103 04:02:59.307365 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" event={"ID":"015d13a2-fc39-4b75-b4ec-8d5bf46c4c00","Type":"ContainerDied","Data":"0592a8755f6a106db2fb1c7ba296474e381859b65e2814d453a87a5362c50e9b"} Jan 03 04:02:59 crc kubenswrapper[4921]: I0103 04:02:59.307424 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0592a8755f6a106db2fb1c7ba296474e381859b65e2814d453a87a5362c50e9b" Jan 03 04:02:59 crc kubenswrapper[4921]: I0103 04:02:59.307444 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-09aa-account-create-qmd7n" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.811574 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-lw567"] Jan 03 04:03:00 crc kubenswrapper[4921]: E0103 04:03:00.812338 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" containerName="mariadb-account-create" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.812355 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" containerName="mariadb-account-create" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.812558 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" containerName="mariadb-account-create" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.813115 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.816022 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.820682 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-z5kw2" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.827315 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lw567"] Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.878758 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqb55\" (UniqueName: \"kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.878972 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.879179 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.979979 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.980069 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqb55\" (UniqueName: \"kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.980156 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.986533 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:00 crc kubenswrapper[4921]: I0103 04:03:00.987132 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:01 crc kubenswrapper[4921]: I0103 04:03:01.014584 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqb55\" (UniqueName: \"kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55\") pod \"glance-db-sync-lw567\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:01 crc kubenswrapper[4921]: I0103 04:03:01.141787 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:01 crc kubenswrapper[4921]: I0103 04:03:01.175383 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:03:01 crc kubenswrapper[4921]: I0103 04:03:01.175464 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:03:01 crc kubenswrapper[4921]: I0103 04:03:01.643405 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lw567"] Jan 03 04:03:01 crc kubenswrapper[4921]: W0103 04:03:01.653405 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda783c4d2_f5e1_4b54_bd67_e99180296cbc.slice/crio-778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896 WatchSource:0}: Error finding container 778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896: Status 404 returned error can't find the container with id 778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896 Jan 03 04:03:02 crc kubenswrapper[4921]: I0103 04:03:02.336125 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lw567" event={"ID":"a783c4d2-f5e1-4b54-bd67-e99180296cbc","Type":"ContainerStarted","Data":"34411416145a1f7bb1308a53deffd0d5c974995fadccc0f1cbd653f54578633e"} Jan 03 04:03:02 crc kubenswrapper[4921]: I0103 04:03:02.336509 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lw567" event={"ID":"a783c4d2-f5e1-4b54-bd67-e99180296cbc","Type":"ContainerStarted","Data":"778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896"} Jan 03 04:03:02 crc kubenswrapper[4921]: I0103 04:03:02.350248 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-lw567" podStartSLOduration=2.350230491 podStartE2EDuration="2.350230491s" podCreationTimestamp="2026-01-03 04:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:03:02.349610074 +0000 UTC m=+1317.961036898" watchObservedRunningTime="2026-01-03 04:03:02.350230491 +0000 UTC m=+1317.961657315" Jan 03 04:03:05 crc kubenswrapper[4921]: I0103 04:03:05.366735 4921 generic.go:334] "Generic (PLEG): container finished" podID="a783c4d2-f5e1-4b54-bd67-e99180296cbc" containerID="34411416145a1f7bb1308a53deffd0d5c974995fadccc0f1cbd653f54578633e" exitCode=0 Jan 03 04:03:05 crc kubenswrapper[4921]: I0103 04:03:05.366794 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lw567" event={"ID":"a783c4d2-f5e1-4b54-bd67-e99180296cbc","Type":"ContainerDied","Data":"34411416145a1f7bb1308a53deffd0d5c974995fadccc0f1cbd653f54578633e"} Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.170228 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.211711 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqb55\" (UniqueName: \"kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55\") pod \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.211983 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data\") pod \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.212137 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data\") pod \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\" (UID: \"a783c4d2-f5e1-4b54-bd67-e99180296cbc\") " Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.222932 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a783c4d2-f5e1-4b54-bd67-e99180296cbc" (UID: "a783c4d2-f5e1-4b54-bd67-e99180296cbc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.230109 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55" (OuterVolumeSpecName: "kube-api-access-rqb55") pod "a783c4d2-f5e1-4b54-bd67-e99180296cbc" (UID: "a783c4d2-f5e1-4b54-bd67-e99180296cbc"). InnerVolumeSpecName "kube-api-access-rqb55". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.270186 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data" (OuterVolumeSpecName: "config-data") pod "a783c4d2-f5e1-4b54-bd67-e99180296cbc" (UID: "a783c4d2-f5e1-4b54-bd67-e99180296cbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.313773 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqb55\" (UniqueName: \"kubernetes.io/projected/a783c4d2-f5e1-4b54-bd67-e99180296cbc-kube-api-access-rqb55\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.313807 4921 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.313819 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a783c4d2-f5e1-4b54-bd67-e99180296cbc-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.383676 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-lw567" event={"ID":"a783c4d2-f5e1-4b54-bd67-e99180296cbc","Type":"ContainerDied","Data":"778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896"} Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.383721 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="778cb85e0f13b251a23ec50597511401f75a71668557c460144b31ae8f33e896" Jan 03 04:03:07 crc kubenswrapper[4921]: I0103 04:03:07.383784 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-lw567" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.671461 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Jan 03 04:03:08 crc kubenswrapper[4921]: E0103 04:03:08.672252 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a783c4d2-f5e1-4b54-bd67-e99180296cbc" containerName="glance-db-sync" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.672294 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="a783c4d2-f5e1-4b54-bd67-e99180296cbc" containerName="glance-db-sync" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.672516 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="a783c4d2-f5e1-4b54-bd67-e99180296cbc" containerName="glance-db-sync" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.674082 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.676572 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.677028 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.677439 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-z5kw2" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.689804 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734546 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734604 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734643 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734673 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbvhw\" (UniqueName: \"kubernetes.io/projected/8a534b6a-d5ee-479c-b105-de684ca6fa62-kube-api-access-tbvhw\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734757 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734784 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734818 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734847 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-logs\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734934 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-dev\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734965 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-sys\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.734995 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.735157 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.735215 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.735333 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.761166 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.762819 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.764391 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.787910 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836532 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836568 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbvhw\" (UniqueName: \"kubernetes.io/projected/8a534b6a-d5ee-479c-b105-de684ca6fa62-kube-api-access-tbvhw\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836604 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836866 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836744 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.836683 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837394 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837447 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837494 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837521 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-logs\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837781 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.837931 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a534b6a-d5ee-479c-b105-de684ca6fa62-logs\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849347 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-dev\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849392 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-sys\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849425 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849513 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849530 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849557 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-sys\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849616 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849637 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-dev\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849672 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849710 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849733 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849815 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-run\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849846 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8a534b6a-d5ee-479c-b105-de684ca6fa62-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.849893 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") device mount path \"/mnt/openstack/pv02\"" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.852854 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbvhw\" (UniqueName: \"kubernetes.io/projected/8a534b6a-d5ee-479c-b105-de684ca6fa62-kube-api-access-tbvhw\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.854512 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.854580 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a534b6a-d5ee-479c-b105-de684ca6fa62-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.857037 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.868850 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"8a534b6a-d5ee-479c-b105-de684ca6fa62\") " pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.951393 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n8rs\" (UniqueName: \"kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.951458 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.951595 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.951898 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.951937 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952013 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952069 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952095 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952163 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952193 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952217 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952249 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952321 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952349 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.952538 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.970848 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:08 crc kubenswrapper[4921]: I0103 04:03:08.988037 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053060 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053131 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053157 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053182 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053210 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053255 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053304 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053338 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n8rs\" (UniqueName: \"kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053358 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053376 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053405 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053430 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053472 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053539 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053579 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053611 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053621 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053659 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053900 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053944 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.053977 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.054018 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.054102 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.057560 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.065554 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.087788 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n8rs\" (UniqueName: \"kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.094341 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.375759 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.465635 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Jan 03 04:03:09 crc kubenswrapper[4921]: W0103 04:03:09.476448 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a534b6a_d5ee_479c_b105_de684ca6fa62.slice/crio-529793b6555669abc95c3e40b25f1d1eaf0be735d9809cb479e498affe60099b WatchSource:0}: Error finding container 529793b6555669abc95c3e40b25f1d1eaf0be735d9809cb479e498affe60099b: Status 404 returned error can't find the container with id 529793b6555669abc95c3e40b25f1d1eaf0be735d9809cb479e498affe60099b Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.673032 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:09 crc kubenswrapper[4921]: I0103 04:03:09.714596 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.408382 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerStarted","Data":"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.408904 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerStarted","Data":"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.408919 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerStarted","Data":"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.408931 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerStarted","Data":"483c14f3d3650e70d45ed8acffbfc37c88cf39159ccb62d61209ef58173d4323"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.409078 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-log" containerID="cri-o://4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" gracePeriod=30 Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.409674 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-api" containerID="cri-o://4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" gracePeriod=30 Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.409741 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-httpd" containerID="cri-o://946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" gracePeriod=30 Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.420693 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8a534b6a-d5ee-479c-b105-de684ca6fa62","Type":"ContainerStarted","Data":"af637d969d2192f4c0e0001a2aae4c53b3e89dd174cd273836b1ef61ea028508"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.420759 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8a534b6a-d5ee-479c-b105-de684ca6fa62","Type":"ContainerStarted","Data":"126d0d1b370d2678f93f50b6aa3390d8d1a84cff2b237e87d5726cfaa0924657"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.420773 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8a534b6a-d5ee-479c-b105-de684ca6fa62","Type":"ContainerStarted","Data":"975e4e64d6479bd21f633f10758512c934f2b8e7cd6ceb5b4241dafc7ef5e76d"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.420788 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"8a534b6a-d5ee-479c-b105-de684ca6fa62","Type":"ContainerStarted","Data":"529793b6555669abc95c3e40b25f1d1eaf0be735d9809cb479e498affe60099b"} Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.444100 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.444081592 podStartE2EDuration="3.444081592s" podCreationTimestamp="2026-01-03 04:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:03:10.441148638 +0000 UTC m=+1326.052575502" watchObservedRunningTime="2026-01-03 04:03:10.444081592 +0000 UTC m=+1326.055508426" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.493485 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.4934656840000002 podStartE2EDuration="2.493465684s" podCreationTimestamp="2026-01-03 04:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:03:10.489594062 +0000 UTC m=+1326.101020906" watchObservedRunningTime="2026-01-03 04:03:10.493465684 +0000 UTC m=+1326.104892518" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.785254 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.880721 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n8rs\" (UniqueName: \"kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.881986 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882031 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882047 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882062 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882082 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882100 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882124 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882141 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882164 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882186 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882206 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882227 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882244 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev\") pod \"95fde5e9-ce6f-490b-b931-9bba0129c037\" (UID: \"95fde5e9-ce6f-490b-b931-9bba0129c037\") " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882500 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev" (OuterVolumeSpecName: "dev") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.882736 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs" (OuterVolumeSpecName: "logs") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889331 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs" (OuterVolumeSpecName: "kube-api-access-2n8rs") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "kube-api-access-2n8rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889397 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys" (OuterVolumeSpecName: "sys") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889513 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889662 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run" (OuterVolumeSpecName: "run") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889706 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889732 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889749 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.889755 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.891858 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.892336 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.893353 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts" (OuterVolumeSpecName: "scripts") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984143 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984199 4921 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-var-locks-brick\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984213 4921 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-lib-modules\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984226 4921 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-dev\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984237 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n8rs\" (UniqueName: \"kubernetes.io/projected/95fde5e9-ce6f-490b-b931-9bba0129c037-kube-api-access-2n8rs\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984248 4921 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-logs\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984259 4921 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-nvme\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984297 4921 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984309 4921 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95fde5e9-ce6f-490b-b931-9bba0129c037-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984327 4921 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-etc-iscsi\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984337 4921 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-sys\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984348 4921 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-scripts\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:10 crc kubenswrapper[4921]: I0103 04:03:10.984358 4921 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/95fde5e9-ce6f-490b-b931-9bba0129c037-run\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.002489 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.005261 4921 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.009541 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data" (OuterVolumeSpecName: "config-data") pod "95fde5e9-ce6f-490b-b931-9bba0129c037" (UID: "95fde5e9-ce6f-490b-b931-9bba0129c037"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.085011 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.085047 4921 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95fde5e9-ce6f-490b-b931-9bba0129c037-config-data\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.085061 4921 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.430678 4921 generic.go:334] "Generic (PLEG): container finished" podID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" exitCode=143 Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.430963 4921 generic.go:334] "Generic (PLEG): container finished" podID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" exitCode=143 Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.430973 4921 generic.go:334] "Generic (PLEG): container finished" podID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" exitCode=143 Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.430772 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerDied","Data":"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005"} Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.431045 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerDied","Data":"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5"} Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.431072 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerDied","Data":"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a"} Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.431086 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"95fde5e9-ce6f-490b-b931-9bba0129c037","Type":"ContainerDied","Data":"483c14f3d3650e70d45ed8acffbfc37c88cf39159ccb62d61209ef58173d4323"} Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.430783 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.431108 4921 scope.go:117] "RemoveContainer" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.463234 4921 scope.go:117] "RemoveContainer" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.466888 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.473291 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.489740 4921 scope.go:117] "RemoveContainer" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.490962 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.491235 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-httpd" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491252 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-httpd" Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.491289 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-api" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491299 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-api" Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.491319 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-log" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491326 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-log" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491439 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-log" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491450 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-api" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.491461 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" containerName="glance-httpd" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.492636 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.495582 4921 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.521921 4921 scope.go:117] "RemoveContainer" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.524546 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": container with ID starting with 4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005 not found: ID does not exist" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.524590 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005"} err="failed to get container status \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": rpc error: code = NotFound desc = could not find container \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": container with ID starting with 4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.524618 4921 scope.go:117] "RemoveContainer" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.525073 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": container with ID starting with 946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5 not found: ID does not exist" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.525095 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5"} err="failed to get container status \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": rpc error: code = NotFound desc = could not find container \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": container with ID starting with 946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.525112 4921 scope.go:117] "RemoveContainer" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" Jan 03 04:03:11 crc kubenswrapper[4921]: E0103 04:03:11.525655 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": container with ID starting with 4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a not found: ID does not exist" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.525708 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a"} err="failed to get container status \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": rpc error: code = NotFound desc = could not find container \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": container with ID starting with 4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.525742 4921 scope.go:117] "RemoveContainer" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526033 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005"} err="failed to get container status \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": rpc error: code = NotFound desc = could not find container \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": container with ID starting with 4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526054 4921 scope.go:117] "RemoveContainer" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526249 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5"} err="failed to get container status \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": rpc error: code = NotFound desc = could not find container \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": container with ID starting with 946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526285 4921 scope.go:117] "RemoveContainer" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526674 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a"} err="failed to get container status \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": rpc error: code = NotFound desc = could not find container \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": container with ID starting with 4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526692 4921 scope.go:117] "RemoveContainer" containerID="4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526924 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005"} err="failed to get container status \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": rpc error: code = NotFound desc = could not find container \"4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005\": container with ID starting with 4618237bf1a2d8fe391e305cae5b5bb6be13de47016b80e5883d9dd3f7961005 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.526955 4921 scope.go:117] "RemoveContainer" containerID="946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.527181 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5"} err="failed to get container status \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": rpc error: code = NotFound desc = could not find container \"946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5\": container with ID starting with 946b0fd2025e0603da8589205a39ec35324ef3910009d5f67d3f9e6e41b252a5 not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.527202 4921 scope.go:117] "RemoveContainer" containerID="4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.531500 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a"} err="failed to get container status \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": rpc error: code = NotFound desc = could not find container \"4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a\": container with ID starting with 4d79261ee5735e14f744b2f5e32e09163018356bbfaf0e215014774205d55f3a not found: ID does not exist" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.555340 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.693953 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694077 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-logs\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694130 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694183 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694241 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694320 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694366 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694393 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694423 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-sys\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694439 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694468 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694486 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-dev\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694534 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.694554 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2nxg\" (UniqueName: \"kubernetes.io/projected/d48a3402-3b34-4080-bc8f-f1ad53021216-kube-api-access-g2nxg\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795855 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-logs\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795908 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795933 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795950 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795976 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.795993 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796017 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796040 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-sys\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796056 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796084 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796099 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-dev\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796114 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796165 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796131 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796171 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796192 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796219 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796234 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796261 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-sys\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796331 4921 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796371 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d48a3402-3b34-4080-bc8f-f1ad53021216-dev\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796774 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.796831 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2nxg\" (UniqueName: \"kubernetes.io/projected/d48a3402-3b34-4080-bc8f-f1ad53021216-kube-api-access-g2nxg\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.797040 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.801445 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.801459 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d48a3402-3b34-4080-bc8f-f1ad53021216-logs\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.814253 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d48a3402-3b34-4080-bc8f-f1ad53021216-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.821487 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.825828 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2nxg\" (UniqueName: \"kubernetes.io/projected/d48a3402-3b34-4080-bc8f-f1ad53021216-kube-api-access-g2nxg\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:11 crc kubenswrapper[4921]: I0103 04:03:11.830722 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d48a3402-3b34-4080-bc8f-f1ad53021216\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:12 crc kubenswrapper[4921]: I0103 04:03:12.113563 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:12 crc kubenswrapper[4921]: I0103 04:03:12.447393 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Jan 03 04:03:12 crc kubenswrapper[4921]: W0103 04:03:12.452739 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd48a3402_3b34_4080_bc8f_f1ad53021216.slice/crio-0f408579a0332ea558fbc1ff9f5a099768cdca1d9ac400eeb57a6a251d17e221 WatchSource:0}: Error finding container 0f408579a0332ea558fbc1ff9f5a099768cdca1d9ac400eeb57a6a251d17e221: Status 404 returned error can't find the container with id 0f408579a0332ea558fbc1ff9f5a099768cdca1d9ac400eeb57a6a251d17e221 Jan 03 04:03:12 crc kubenswrapper[4921]: I0103 04:03:12.902097 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95fde5e9-ce6f-490b-b931-9bba0129c037" path="/var/lib/kubelet/pods/95fde5e9-ce6f-490b-b931-9bba0129c037/volumes" Jan 03 04:03:13 crc kubenswrapper[4921]: I0103 04:03:13.464238 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d48a3402-3b34-4080-bc8f-f1ad53021216","Type":"ContainerStarted","Data":"fab1e83466f81d9b30e0b54eb94fc7a524f6ef536c123b804c26a90c0e5c7df9"} Jan 03 04:03:13 crc kubenswrapper[4921]: I0103 04:03:13.464673 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d48a3402-3b34-4080-bc8f-f1ad53021216","Type":"ContainerStarted","Data":"b095ce8f4f45b66f7e5a2da53f55907a4016e183b4eada0d20127d2e3201359f"} Jan 03 04:03:13 crc kubenswrapper[4921]: I0103 04:03:13.464683 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d48a3402-3b34-4080-bc8f-f1ad53021216","Type":"ContainerStarted","Data":"9a6290e830f094ea45485b0afcea0e52cac110f533b36632ead87f7d1170bd09"} Jan 03 04:03:13 crc kubenswrapper[4921]: I0103 04:03:13.464695 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"d48a3402-3b34-4080-bc8f-f1ad53021216","Type":"ContainerStarted","Data":"0f408579a0332ea558fbc1ff9f5a099768cdca1d9ac400eeb57a6a251d17e221"} Jan 03 04:03:13 crc kubenswrapper[4921]: I0103 04:03:13.506893 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.506870419 podStartE2EDuration="2.506870419s" podCreationTimestamp="2026-01-03 04:03:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-03 04:03:13.495698297 +0000 UTC m=+1329.107125131" watchObservedRunningTime="2026-01-03 04:03:13.506870419 +0000 UTC m=+1329.118297253" Jan 03 04:03:18 crc kubenswrapper[4921]: I0103 04:03:18.988628 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:18 crc kubenswrapper[4921]: I0103 04:03:18.989352 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:18 crc kubenswrapper[4921]: I0103 04:03:18.989386 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.027636 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.028566 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.062131 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.527535 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.527613 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.527643 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.547041 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.547469 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:19 crc kubenswrapper[4921]: I0103 04:03:19.552917 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.114074 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.114369 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.114382 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.142072 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.147831 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.160883 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.551107 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.551162 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.551175 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.563173 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.564882 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:22 crc kubenswrapper[4921]: I0103 04:03:22.565929 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Jan 03 04:03:31 crc kubenswrapper[4921]: I0103 04:03:31.176039 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:03:31 crc kubenswrapper[4921]: I0103 04:03:31.176650 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.176100 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.176581 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.176624 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.177165 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.177217 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3" gracePeriod=600 Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.909193 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3" exitCode=0 Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.909281 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3"} Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.909657 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f"} Jan 03 04:04:01 crc kubenswrapper[4921]: I0103 04:04:01.909681 4921 scope.go:117] "RemoveContainer" containerID="a38dcfe4c7faa79da03f79691fbcda0d46c3d4c23798ddf41d926baf7635eee4" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.128400 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.132059 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.146906 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.278153 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pjnm\" (UniqueName: \"kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.278238 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.278315 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.380144 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pjnm\" (UniqueName: \"kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.380219 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.380294 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.380902 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.380914 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.401234 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pjnm\" (UniqueName: \"kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm\") pod \"certified-operators-srs66\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.455492 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:43 crc kubenswrapper[4921]: I0103 04:04:43.761064 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:44 crc kubenswrapper[4921]: I0103 04:04:44.527661 4921 generic.go:334] "Generic (PLEG): container finished" podID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerID="4760db133a6b0db60e647b588cc35b40f1ffc9534331e10f79ce57c174ef6a25" exitCode=0 Jan 03 04:04:44 crc kubenswrapper[4921]: I0103 04:04:44.527871 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerDied","Data":"4760db133a6b0db60e647b588cc35b40f1ffc9534331e10f79ce57c174ef6a25"} Jan 03 04:04:44 crc kubenswrapper[4921]: I0103 04:04:44.527897 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerStarted","Data":"17151cc852d4280977ac208b5875d9303c81e2d08e9704738f06e019dc3b04c5"} Jan 03 04:04:44 crc kubenswrapper[4921]: I0103 04:04:44.529992 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 04:04:45 crc kubenswrapper[4921]: I0103 04:04:45.541164 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerStarted","Data":"954d69c7d1955c8ef1ef4d092b0670524878702b0778a822e6ca7497125c3c44"} Jan 03 04:04:46 crc kubenswrapper[4921]: I0103 04:04:46.551428 4921 generic.go:334] "Generic (PLEG): container finished" podID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerID="954d69c7d1955c8ef1ef4d092b0670524878702b0778a822e6ca7497125c3c44" exitCode=0 Jan 03 04:04:46 crc kubenswrapper[4921]: I0103 04:04:46.551511 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerDied","Data":"954d69c7d1955c8ef1ef4d092b0670524878702b0778a822e6ca7497125c3c44"} Jan 03 04:04:47 crc kubenswrapper[4921]: I0103 04:04:47.560669 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerStarted","Data":"d2cda3bc2ffcb2efca6172e079b255645f76c9fb8ec8cbd5657c3fb8059771a3"} Jan 03 04:04:47 crc kubenswrapper[4921]: I0103 04:04:47.578966 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-srs66" podStartSLOduration=2.093894994 podStartE2EDuration="4.578942475s" podCreationTimestamp="2026-01-03 04:04:43 +0000 UTC" firstStartedPulling="2026-01-03 04:04:44.529641155 +0000 UTC m=+1420.141067979" lastFinishedPulling="2026-01-03 04:04:47.014688626 +0000 UTC m=+1422.626115460" observedRunningTime="2026-01-03 04:04:47.574947438 +0000 UTC m=+1423.186374282" watchObservedRunningTime="2026-01-03 04:04:47.578942475 +0000 UTC m=+1423.190369309" Jan 03 04:04:53 crc kubenswrapper[4921]: I0103 04:04:53.456449 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:53 crc kubenswrapper[4921]: I0103 04:04:53.456907 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:53 crc kubenswrapper[4921]: I0103 04:04:53.526391 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:53 crc kubenswrapper[4921]: I0103 04:04:53.695068 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:53 crc kubenswrapper[4921]: I0103 04:04:53.766433 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:55 crc kubenswrapper[4921]: I0103 04:04:55.664317 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-srs66" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="registry-server" containerID="cri-o://d2cda3bc2ffcb2efca6172e079b255645f76c9fb8ec8cbd5657c3fb8059771a3" gracePeriod=2 Jan 03 04:04:56 crc kubenswrapper[4921]: I0103 04:04:56.696186 4921 generic.go:334] "Generic (PLEG): container finished" podID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerID="d2cda3bc2ffcb2efca6172e079b255645f76c9fb8ec8cbd5657c3fb8059771a3" exitCode=0 Jan 03 04:04:56 crc kubenswrapper[4921]: I0103 04:04:56.696372 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerDied","Data":"d2cda3bc2ffcb2efca6172e079b255645f76c9fb8ec8cbd5657c3fb8059771a3"} Jan 03 04:04:56 crc kubenswrapper[4921]: I0103 04:04:56.964307 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.114577 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content\") pod \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.114676 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pjnm\" (UniqueName: \"kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm\") pod \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.114716 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities\") pod \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\" (UID: \"b2ef78d9-a5fa-415b-8990-c1cf05e940c3\") " Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.115848 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities" (OuterVolumeSpecName: "utilities") pod "b2ef78d9-a5fa-415b-8990-c1cf05e940c3" (UID: "b2ef78d9-a5fa-415b-8990-c1cf05e940c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.122766 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm" (OuterVolumeSpecName: "kube-api-access-9pjnm") pod "b2ef78d9-a5fa-415b-8990-c1cf05e940c3" (UID: "b2ef78d9-a5fa-415b-8990-c1cf05e940c3"). InnerVolumeSpecName "kube-api-access-9pjnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.206697 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2ef78d9-a5fa-415b-8990-c1cf05e940c3" (UID: "b2ef78d9-a5fa-415b-8990-c1cf05e940c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.216182 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pjnm\" (UniqueName: \"kubernetes.io/projected/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-kube-api-access-9pjnm\") on node \"crc\" DevicePath \"\"" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.216331 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.216410 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2ef78d9-a5fa-415b-8990-c1cf05e940c3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.709967 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-srs66" event={"ID":"b2ef78d9-a5fa-415b-8990-c1cf05e940c3","Type":"ContainerDied","Data":"17151cc852d4280977ac208b5875d9303c81e2d08e9704738f06e019dc3b04c5"} Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.710016 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-srs66" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.710029 4921 scope.go:117] "RemoveContainer" containerID="d2cda3bc2ffcb2efca6172e079b255645f76c9fb8ec8cbd5657c3fb8059771a3" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.751737 4921 scope.go:117] "RemoveContainer" containerID="954d69c7d1955c8ef1ef4d092b0670524878702b0778a822e6ca7497125c3c44" Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.757236 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.766463 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-srs66"] Jan 03 04:04:57 crc kubenswrapper[4921]: I0103 04:04:57.774377 4921 scope.go:117] "RemoveContainer" containerID="4760db133a6b0db60e647b588cc35b40f1ffc9534331e10f79ce57c174ef6a25" Jan 03 04:04:58 crc kubenswrapper[4921]: I0103 04:04:58.895110 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" path="/var/lib/kubelet/pods/b2ef78d9-a5fa-415b-8990-c1cf05e940c3/volumes" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.910908 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:04:59 crc kubenswrapper[4921]: E0103 04:04:59.911354 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="extract-utilities" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.911377 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="extract-utilities" Jan 03 04:04:59 crc kubenswrapper[4921]: E0103 04:04:59.911420 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="registry-server" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.911432 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="registry-server" Jan 03 04:04:59 crc kubenswrapper[4921]: E0103 04:04:59.911477 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="extract-content" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.911489 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="extract-content" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.911748 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2ef78d9-a5fa-415b-8990-c1cf05e940c3" containerName="registry-server" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.913424 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:04:59 crc kubenswrapper[4921]: I0103 04:04:59.939187 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.070534 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.070599 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrrt\" (UniqueName: \"kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.070788 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.172686 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.172766 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.172814 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrrt\" (UniqueName: \"kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.173325 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.173679 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.204106 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrrt\" (UniqueName: \"kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt\") pod \"community-operators-jxdhn\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.280929 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.573554 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:05:00 crc kubenswrapper[4921]: I0103 04:05:00.736742 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerStarted","Data":"b613b9e99768fa941adcfd302b958e0e16ce3701cf8f9ba9571c5ba58dba6002"} Jan 03 04:05:01 crc kubenswrapper[4921]: I0103 04:05:01.747100 4921 generic.go:334] "Generic (PLEG): container finished" podID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerID="88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1" exitCode=0 Jan 03 04:05:01 crc kubenswrapper[4921]: I0103 04:05:01.747140 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerDied","Data":"88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1"} Jan 03 04:05:03 crc kubenswrapper[4921]: I0103 04:05:03.770442 4921 generic.go:334] "Generic (PLEG): container finished" podID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerID="118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85" exitCode=0 Jan 03 04:05:03 crc kubenswrapper[4921]: I0103 04:05:03.770517 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerDied","Data":"118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85"} Jan 03 04:05:04 crc kubenswrapper[4921]: I0103 04:05:04.792636 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerStarted","Data":"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a"} Jan 03 04:05:04 crc kubenswrapper[4921]: I0103 04:05:04.823091 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jxdhn" podStartSLOduration=3.367346501 podStartE2EDuration="5.823059984s" podCreationTimestamp="2026-01-03 04:04:59 +0000 UTC" firstStartedPulling="2026-01-03 04:05:01.748755493 +0000 UTC m=+1437.360182337" lastFinishedPulling="2026-01-03 04:05:04.204468996 +0000 UTC m=+1439.815895820" observedRunningTime="2026-01-03 04:05:04.81200996 +0000 UTC m=+1440.423436794" watchObservedRunningTime="2026-01-03 04:05:04.823059984 +0000 UTC m=+1440.434486848" Jan 03 04:05:10 crc kubenswrapper[4921]: I0103 04:05:10.282028 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:10 crc kubenswrapper[4921]: I0103 04:05:10.282506 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:10 crc kubenswrapper[4921]: I0103 04:05:10.352925 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:10 crc kubenswrapper[4921]: I0103 04:05:10.912723 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:10 crc kubenswrapper[4921]: I0103 04:05:10.982390 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:05:12 crc kubenswrapper[4921]: I0103 04:05:12.863061 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jxdhn" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="registry-server" containerID="cri-o://5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a" gracePeriod=2 Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.332371 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.403846 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twrrt\" (UniqueName: \"kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt\") pod \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.403942 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content\") pod \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.404103 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities\") pod \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\" (UID: \"b1f2230a-d17c-4744-a90d-5e75cd01e24f\") " Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.405303 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities" (OuterVolumeSpecName: "utilities") pod "b1f2230a-d17c-4744-a90d-5e75cd01e24f" (UID: "b1f2230a-d17c-4744-a90d-5e75cd01e24f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.409097 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt" (OuterVolumeSpecName: "kube-api-access-twrrt") pod "b1f2230a-d17c-4744-a90d-5e75cd01e24f" (UID: "b1f2230a-d17c-4744-a90d-5e75cd01e24f"). InnerVolumeSpecName "kube-api-access-twrrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.505644 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twrrt\" (UniqueName: \"kubernetes.io/projected/b1f2230a-d17c-4744-a90d-5e75cd01e24f-kube-api-access-twrrt\") on node \"crc\" DevicePath \"\"" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.505689 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.596414 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1f2230a-d17c-4744-a90d-5e75cd01e24f" (UID: "b1f2230a-d17c-4744-a90d-5e75cd01e24f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.607605 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1f2230a-d17c-4744-a90d-5e75cd01e24f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.878717 4921 generic.go:334] "Generic (PLEG): container finished" podID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerID="5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a" exitCode=0 Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.878798 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerDied","Data":"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a"} Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.878812 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxdhn" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.878861 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxdhn" event={"ID":"b1f2230a-d17c-4744-a90d-5e75cd01e24f","Type":"ContainerDied","Data":"b613b9e99768fa941adcfd302b958e0e16ce3701cf8f9ba9571c5ba58dba6002"} Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.878901 4921 scope.go:117] "RemoveContainer" containerID="5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.916499 4921 scope.go:117] "RemoveContainer" containerID="118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.941920 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.954079 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jxdhn"] Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.961089 4921 scope.go:117] "RemoveContainer" containerID="88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.994340 4921 scope.go:117] "RemoveContainer" containerID="5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a" Jan 03 04:05:13 crc kubenswrapper[4921]: E0103 04:05:13.994866 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a\": container with ID starting with 5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a not found: ID does not exist" containerID="5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.994957 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a"} err="failed to get container status \"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a\": rpc error: code = NotFound desc = could not find container \"5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a\": container with ID starting with 5619b89e98478143db3cf9963bfc393ca66c344a29ec0a17f0abb1d516e07b5a not found: ID does not exist" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.994997 4921 scope.go:117] "RemoveContainer" containerID="118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85" Jan 03 04:05:13 crc kubenswrapper[4921]: E0103 04:05:13.995427 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85\": container with ID starting with 118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85 not found: ID does not exist" containerID="118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.995458 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85"} err="failed to get container status \"118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85\": rpc error: code = NotFound desc = could not find container \"118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85\": container with ID starting with 118961ffea1a616c26d7c83c499b0c1510011545072cacba1205482568881e85 not found: ID does not exist" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.995483 4921 scope.go:117] "RemoveContainer" containerID="88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1" Jan 03 04:05:13 crc kubenswrapper[4921]: E0103 04:05:13.995845 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1\": container with ID starting with 88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1 not found: ID does not exist" containerID="88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1" Jan 03 04:05:13 crc kubenswrapper[4921]: I0103 04:05:13.995867 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1"} err="failed to get container status \"88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1\": rpc error: code = NotFound desc = could not find container \"88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1\": container with ID starting with 88aae8726d2a7543e50e87645e8aca632f1803aaa0914bad39b2293604adb6f1 not found: ID does not exist" Jan 03 04:05:14 crc kubenswrapper[4921]: I0103 04:05:14.898932 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" path="/var/lib/kubelet/pods/b1f2230a-d17c-4744-a90d-5e75cd01e24f/volumes" Jan 03 04:06:01 crc kubenswrapper[4921]: I0103 04:06:01.175867 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:06:01 crc kubenswrapper[4921]: I0103 04:06:01.176556 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:06:31 crc kubenswrapper[4921]: I0103 04:06:31.175826 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:06:31 crc kubenswrapper[4921]: I0103 04:06:31.176398 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.176539 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.177084 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.177127 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.177716 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.177770 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" gracePeriod=600 Jan 03 04:07:01 crc kubenswrapper[4921]: E0103 04:07:01.303418 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.891407 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" exitCode=0 Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.891506 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f"} Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.891758 4921 scope.go:117] "RemoveContainer" containerID="39da3dbf8c9e26abbe9c867056fea5a92be228065dc7283b3fdce470e255f2f3" Jan 03 04:07:01 crc kubenswrapper[4921]: I0103 04:07:01.892325 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:07:01 crc kubenswrapper[4921]: E0103 04:07:01.892593 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:07:14 crc kubenswrapper[4921]: I0103 04:07:14.894642 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:07:14 crc kubenswrapper[4921]: E0103 04:07:14.895808 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:07:16 crc kubenswrapper[4921]: I0103 04:07:16.438398 4921 scope.go:117] "RemoveContainer" containerID="62da10a9f4cd1fb2a083feaa6a9a63b9f43e5c16064bd03065dfb82818a1502b" Jan 03 04:07:16 crc kubenswrapper[4921]: I0103 04:07:16.464469 4921 scope.go:117] "RemoveContainer" containerID="a35139db0766dc14a29869569638c50cf2adba17eb3a5cbee63ca5128426f596" Jan 03 04:07:28 crc kubenswrapper[4921]: I0103 04:07:28.884169 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:07:28 crc kubenswrapper[4921]: E0103 04:07:28.886083 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:07:40 crc kubenswrapper[4921]: I0103 04:07:40.885062 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:07:40 crc kubenswrapper[4921]: E0103 04:07:40.886007 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:07:55 crc kubenswrapper[4921]: I0103 04:07:55.883575 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:07:55 crc kubenswrapper[4921]: E0103 04:07:55.884318 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:08:09 crc kubenswrapper[4921]: I0103 04:08:09.883888 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:08:09 crc kubenswrapper[4921]: E0103 04:08:09.884706 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:08:16 crc kubenswrapper[4921]: I0103 04:08:16.564432 4921 scope.go:117] "RemoveContainer" containerID="1a1d9d007caf27d0bf2c979d1de8879cab94c0ec124f85a95b0d026333bda6b9" Jan 03 04:08:16 crc kubenswrapper[4921]: I0103 04:08:16.626517 4921 scope.go:117] "RemoveContainer" containerID="9b44bb323f503fc1bfcf41778be5c4f7739c8569759db79bbfa72e0003f514c4" Jan 03 04:08:20 crc kubenswrapper[4921]: I0103 04:08:20.885103 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:08:20 crc kubenswrapper[4921]: E0103 04:08:20.885873 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:08:33 crc kubenswrapper[4921]: I0103 04:08:33.883831 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:08:33 crc kubenswrapper[4921]: E0103 04:08:33.885050 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:08:45 crc kubenswrapper[4921]: I0103 04:08:45.884747 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:08:45 crc kubenswrapper[4921]: E0103 04:08:45.885503 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:08:50 crc kubenswrapper[4921]: I0103 04:08:50.061181 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-bj9mx"] Jan 03 04:08:50 crc kubenswrapper[4921]: I0103 04:08:50.068911 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-bj9mx"] Jan 03 04:08:50 crc kubenswrapper[4921]: I0103 04:08:50.898738 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee9d2b0d-28d2-4d99-a67c-71af6d2ee261" path="/var/lib/kubelet/pods/ee9d2b0d-28d2-4d99-a67c-71af6d2ee261/volumes" Jan 03 04:08:59 crc kubenswrapper[4921]: I0103 04:08:59.884324 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:08:59 crc kubenswrapper[4921]: E0103 04:08:59.885218 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:09:01 crc kubenswrapper[4921]: I0103 04:09:01.030480 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-9b09-account-create-5nc6d"] Jan 03 04:09:01 crc kubenswrapper[4921]: I0103 04:09:01.039342 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-9b09-account-create-5nc6d"] Jan 03 04:09:02 crc kubenswrapper[4921]: I0103 04:09:02.898915 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ad2843b-da3a-4404-af04-6b552cddfa6d" path="/var/lib/kubelet/pods/1ad2843b-da3a-4404-af04-6b552cddfa6d/volumes" Jan 03 04:09:10 crc kubenswrapper[4921]: I0103 04:09:10.884579 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:09:10 crc kubenswrapper[4921]: E0103 04:09:10.885371 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:09:16 crc kubenswrapper[4921]: I0103 04:09:16.724254 4921 scope.go:117] "RemoveContainer" containerID="e27ec8fc352ae5fa4499b9400a934e3e933091d3190d62428162ab14e0ddbc36" Jan 03 04:09:16 crc kubenswrapper[4921]: I0103 04:09:16.769745 4921 scope.go:117] "RemoveContainer" containerID="5a0c7beba4b4884779fa1b1f4ae3bc0a21477bcf48d10362755b7f3e6df03e7f" Jan 03 04:09:16 crc kubenswrapper[4921]: I0103 04:09:16.808978 4921 scope.go:117] "RemoveContainer" containerID="b0476718bfde774e30f4630e50fbac430b23e66798d51049f7989a1800f73e81" Jan 03 04:09:20 crc kubenswrapper[4921]: I0103 04:09:20.056150 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-45jwt"] Jan 03 04:09:20 crc kubenswrapper[4921]: I0103 04:09:20.065057 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-45jwt"] Jan 03 04:09:20 crc kubenswrapper[4921]: I0103 04:09:20.897235 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b022c441-4cd8-4f96-a93b-cd194a4a505b" path="/var/lib/kubelet/pods/b022c441-4cd8-4f96-a93b-cd194a4a505b/volumes" Jan 03 04:09:25 crc kubenswrapper[4921]: I0103 04:09:25.884516 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:09:25 crc kubenswrapper[4921]: E0103 04:09:25.885260 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:09:26 crc kubenswrapper[4921]: I0103 04:09:26.032792 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6vmcw"] Jan 03 04:09:26 crc kubenswrapper[4921]: I0103 04:09:26.039678 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6vmcw"] Jan 03 04:09:26 crc kubenswrapper[4921]: I0103 04:09:26.898749 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80" path="/var/lib/kubelet/pods/34f4fd8b-e3c5-40ff-8ebd-39ec9e6cff80/volumes" Jan 03 04:09:40 crc kubenswrapper[4921]: I0103 04:09:40.885013 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:09:40 crc kubenswrapper[4921]: E0103 04:09:40.886163 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:09:55 crc kubenswrapper[4921]: I0103 04:09:55.883419 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:09:55 crc kubenswrapper[4921]: E0103 04:09:55.885347 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:10:09 crc kubenswrapper[4921]: I0103 04:10:09.884349 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:10:09 crc kubenswrapper[4921]: E0103 04:10:09.885183 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:10:16 crc kubenswrapper[4921]: I0103 04:10:16.967084 4921 scope.go:117] "RemoveContainer" containerID="6d6c144736539e6b6cea09efc8064df216c26a5e5837bc194e0ed3984ea93603" Jan 03 04:10:17 crc kubenswrapper[4921]: I0103 04:10:17.026825 4921 scope.go:117] "RemoveContainer" containerID="9f1294aa84cde1c458fb6eeafea7bf173f557e036e7324c31fa60be3884caf6b" Jan 03 04:10:21 crc kubenswrapper[4921]: I0103 04:10:21.883596 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:10:21 crc kubenswrapper[4921]: E0103 04:10:21.884254 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:10:35 crc kubenswrapper[4921]: I0103 04:10:35.883752 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:10:35 crc kubenswrapper[4921]: E0103 04:10:35.884453 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:10:46 crc kubenswrapper[4921]: I0103 04:10:46.884019 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:10:46 crc kubenswrapper[4921]: E0103 04:10:46.885441 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.796744 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-g8xn7/must-gather-x86lq"] Jan 03 04:10:47 crc kubenswrapper[4921]: E0103 04:10:47.797560 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="extract-content" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.797594 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="extract-content" Jan 03 04:10:47 crc kubenswrapper[4921]: E0103 04:10:47.797625 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="extract-utilities" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.797639 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="extract-utilities" Jan 03 04:10:47 crc kubenswrapper[4921]: E0103 04:10:47.797689 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="registry-server" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.797703 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="registry-server" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.797939 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1f2230a-d17c-4744-a90d-5e75cd01e24f" containerName="registry-server" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.799194 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.801795 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-g8xn7"/"openshift-service-ca.crt" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.802427 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-g8xn7"/"kube-root-ca.crt" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.820388 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-g8xn7/must-gather-x86lq"] Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.894592 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzhvh\" (UniqueName: \"kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.894644 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.995870 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzhvh\" (UniqueName: \"kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.995937 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:47 crc kubenswrapper[4921]: I0103 04:10:47.996460 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:48 crc kubenswrapper[4921]: I0103 04:10:48.013859 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzhvh\" (UniqueName: \"kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh\") pod \"must-gather-x86lq\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:48 crc kubenswrapper[4921]: I0103 04:10:48.122885 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:10:48 crc kubenswrapper[4921]: I0103 04:10:48.558983 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-g8xn7/must-gather-x86lq"] Jan 03 04:10:48 crc kubenswrapper[4921]: I0103 04:10:48.562817 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 04:10:48 crc kubenswrapper[4921]: I0103 04:10:48.988352 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g8xn7/must-gather-x86lq" event={"ID":"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9","Type":"ContainerStarted","Data":"a17e33130b53f84cb26fde0a49b86b973ab6f171f0420b45b4c8868d8ae9fc0f"} Jan 03 04:10:53 crc kubenswrapper[4921]: I0103 04:10:53.022246 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g8xn7/must-gather-x86lq" event={"ID":"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9","Type":"ContainerStarted","Data":"29c5b4937cc4fa0759a4eb153bbf02fa2e9fa52a6dfd51db928ef82bac27fd3b"} Jan 03 04:10:53 crc kubenswrapper[4921]: I0103 04:10:53.022913 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g8xn7/must-gather-x86lq" event={"ID":"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9","Type":"ContainerStarted","Data":"09e7bda62f70b51fe5c2be927c44df2555fa7f17a9e2d62b2affb323facd5ad4"} Jan 03 04:10:53 crc kubenswrapper[4921]: I0103 04:10:53.038803 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-g8xn7/must-gather-x86lq" podStartSLOduration=2.19040691 podStartE2EDuration="6.038788367s" podCreationTimestamp="2026-01-03 04:10:47 +0000 UTC" firstStartedPulling="2026-01-03 04:10:48.562750612 +0000 UTC m=+1784.174177446" lastFinishedPulling="2026-01-03 04:10:52.411132079 +0000 UTC m=+1788.022558903" observedRunningTime="2026-01-03 04:10:53.037862102 +0000 UTC m=+1788.649288926" watchObservedRunningTime="2026-01-03 04:10:53.038788367 +0000 UTC m=+1788.650215191" Jan 03 04:10:59 crc kubenswrapper[4921]: I0103 04:10:59.883767 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:10:59 crc kubenswrapper[4921]: E0103 04:10:59.884481 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:11:12 crc kubenswrapper[4921]: I0103 04:11:12.888386 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:11:12 crc kubenswrapper[4921]: E0103 04:11:12.889108 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:11:26 crc kubenswrapper[4921]: I0103 04:11:26.883545 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:11:26 crc kubenswrapper[4921]: E0103 04:11:26.884247 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.598067 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/util/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.744206 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/util/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.750912 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/pull/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.762624 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/pull/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.904523 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/extract/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.934608 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/util/0.log" Jan 03 04:11:30 crc kubenswrapper[4921]: I0103 04:11:30.957535 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1d782433d8f1b50ea1db30fe4c554b57fbd2ee68487392fb880a5d5d4dglr67_70ec3f78-d438-442a-8378-d23a76219ca0/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.100577 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.239923 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.252358 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.271081 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.408537 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.413474 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/extract/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.451095 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5f7257904720cd11a481bcace8074ff5a00a306360575c73c97442bf21t268q_7dfc2c0b-f94b-4735-9ead-47ce29635737/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.582489 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.767694 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.781213 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.785958 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.944433 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/util/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.969173 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/pull/0.log" Jan 03 04:11:31 crc kubenswrapper[4921]: I0103 04:11:31.976907 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_676b439cb92f9f44554843f050413ed1e37cf652cd359df0fdc2f4aeefkqmhr_257464fd-1827-4d32-a0bc-b37c100622a0/extract/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.110314 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/util/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.276097 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/pull/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.419858 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/util/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.444404 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/pull/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.583551 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/pull/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.608220 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/util/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.614844 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_798531b9d9a078af26f5f153dd8093f0980ac32bb052a41050c010ef74qlr6p_a4c5e860-d6d5-4b05-965a-042db7d97a45/extract/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.771173 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/util/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.919410 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/pull/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.928463 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/util/0.log" Jan 03 04:11:32 crc kubenswrapper[4921]: I0103 04:11:32.978085 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.117432 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.124545 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.143681 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590l8mgf_23d84029-73ca-4284-94e4-717bb7c8463d/extract/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.272631 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.478161 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.496244 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.553053 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.655678 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.685044 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/extract/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.685608 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a1916e853cc7b15de58e1b135fc3d4209d9752d32839650491e625e07flmj6k_4ca0e16f-8e0c-455a-9767-951056bd29cb/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.726557 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.924909 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/util/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.928859 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/pull/0.log" Jan 03 04:11:33 crc kubenswrapper[4921]: I0103 04:11:33.928918 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/pull/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.094687 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/util/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.159418 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/extract/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.169555 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f347a50fc1e77e2de54efc3b84d64a5c33eca5195620d0b65d31f525619ws8h_876c356f-d130-4a6f-b888-5a32ab0d83da/pull/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.195366 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-b8df6954c-rpzb7_771c903b-35e6-4f43-9b20-ab43e6726425/kube-rbac-proxy/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.361521 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-b8df6954c-rpzb7_771c903b-35e6-4f43-9b20-ab43e6726425/manager/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.383220 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-vtjh6_994de956-f9e4-4ba0-8b97-a9a719655c63/registry-server/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.420911 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6cb8f6dc8b-vtrl7_75035b13-a1d1-44f4-bc84-2f39ff65f476/kube-rbac-proxy/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.518433 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6cb8f6dc8b-vtrl7_75035b13-a1d1-44f4-bc84-2f39ff65f476/manager/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.575080 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-rjwtj_69216a14-e1e4-4b97-8faf-d858a829143d/registry-server/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.614805 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-c6c695967-t8mx7_f7d4e479-994e-4d8f-ba48-f5c4961a8d0d/kube-rbac-proxy/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.696678 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-c6c695967-t8mx7_f7d4e479-994e-4d8f-ba48-f5c4961a8d0d/manager/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.741918 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-ltmf9_59f09088-e73a-4017-ba74-9e208902a0ac/registry-server/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.762147 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6486d9c577-kpxkb_51565a9f-65ee-4007-b46d-bf3100a220a6/kube-rbac-proxy/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.905935 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-6486d9c577-kpxkb_51565a9f-65ee-4007-b46d-bf3100a220a6/manager/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.924918 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-fjmv9_2486f9b9-492d-47e6-9620-9a1b53d419eb/registry-server/0.log" Jan 03 04:11:34 crc kubenswrapper[4921]: I0103 04:11:34.971520 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-677896c6f-dsgnm_63024a63-5cbc-4f16-af95-cd2c087b42dd/kube-rbac-proxy/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.102175 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-677896c6f-dsgnm_63024a63-5cbc-4f16-af95-cd2c087b42dd/manager/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.115565 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-4cq2f_7d33fc60-b79f-4e6f-94e2-9b54e9e7a8ba/registry-server/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.145845 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-hd77p_bc1fdb74-827f-4dcb-bb5e-0973b8ae66bf/operator/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.283354 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-277lk_69284c8d-7396-4067-adad-72b4e024b197/registry-server/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.322610 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-577f5cc87d-645gp_7afe11a6-6cd5-4779-903b-fa49ae0fb5e5/kube-rbac-proxy/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.411398 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-577f5cc87d-645gp_7afe11a6-6cd5-4779-903b-fa49ae0fb5e5/manager/0.log" Jan 03 04:11:35 crc kubenswrapper[4921]: I0103 04:11:35.500779 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-54hh2_d335cc0f-f8c6-4810-9113-cb8b934d142e/registry-server/0.log" Jan 03 04:11:39 crc kubenswrapper[4921]: I0103 04:11:39.883928 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:11:39 crc kubenswrapper[4921]: E0103 04:11:39.884596 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:11:50 crc kubenswrapper[4921]: I0103 04:11:50.323433 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-fxkcl_8bc7bf28-e5c6-4c6d-9674-78e4892ed7f0/control-plane-machine-set-operator/0.log" Jan 03 04:11:50 crc kubenswrapper[4921]: I0103 04:11:50.461070 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2dn4r_478cd3d6-1d5e-4fe0-a103-2caa602ed743/kube-rbac-proxy/0.log" Jan 03 04:11:50 crc kubenswrapper[4921]: I0103 04:11:50.476564 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-2dn4r_478cd3d6-1d5e-4fe0-a103-2caa602ed743/machine-api-operator/0.log" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.206817 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.208888 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.229344 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.278628 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.278693 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrkdc\" (UniqueName: \"kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.279147 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.380477 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.380845 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.380885 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrkdc\" (UniqueName: \"kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.381187 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.381427 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.400478 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrkdc\" (UniqueName: \"kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc\") pod \"redhat-operators-j772m\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.533487 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:11:53 crc kubenswrapper[4921]: I0103 04:11:53.884480 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:11:53 crc kubenswrapper[4921]: E0103 04:11:53.884774 4921 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-cctxw_openshift-machine-config-operator(429ab47e-68f8-4b60-aa4c-ab79a764b7db)\"" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" Jan 03 04:11:54 crc kubenswrapper[4921]: I0103 04:11:54.005154 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:11:54 crc kubenswrapper[4921]: I0103 04:11:54.513409 4921 generic.go:334] "Generic (PLEG): container finished" podID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerID="f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0" exitCode=0 Jan 03 04:11:54 crc kubenswrapper[4921]: I0103 04:11:54.513464 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerDied","Data":"f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0"} Jan 03 04:11:54 crc kubenswrapper[4921]: I0103 04:11:54.513707 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerStarted","Data":"1338e8b9083016002a017b59970fa715fe9a7b1093a683f014d0ffe2f2a2936a"} Jan 03 04:11:55 crc kubenswrapper[4921]: I0103 04:11:55.521008 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerStarted","Data":"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7"} Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.389499 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.391206 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.403642 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.430311 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.430381 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj2g4\" (UniqueName: \"kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.430419 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.532069 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.532206 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.532239 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj2g4\" (UniqueName: \"kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.533028 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.533444 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.553002 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj2g4\" (UniqueName: \"kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4\") pod \"redhat-marketplace-s2w2g\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:56 crc kubenswrapper[4921]: I0103 04:11:56.707263 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:11:57 crc kubenswrapper[4921]: I0103 04:11:57.143861 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:11:57 crc kubenswrapper[4921]: W0103 04:11:57.147923 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5939ae6_1c71_4953_831f_604ea8dccbef.slice/crio-71531f6245d4868df60d01cb75e7779dba60fcf7a53ec32d9d275b58259f3f86 WatchSource:0}: Error finding container 71531f6245d4868df60d01cb75e7779dba60fcf7a53ec32d9d275b58259f3f86: Status 404 returned error can't find the container with id 71531f6245d4868df60d01cb75e7779dba60fcf7a53ec32d9d275b58259f3f86 Jan 03 04:11:57 crc kubenswrapper[4921]: I0103 04:11:57.540438 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerStarted","Data":"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076"} Jan 03 04:11:57 crc kubenswrapper[4921]: I0103 04:11:57.540892 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerStarted","Data":"71531f6245d4868df60d01cb75e7779dba60fcf7a53ec32d9d275b58259f3f86"} Jan 03 04:11:57 crc kubenswrapper[4921]: I0103 04:11:57.542980 4921 generic.go:334] "Generic (PLEG): container finished" podID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerID="7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7" exitCode=0 Jan 03 04:11:57 crc kubenswrapper[4921]: I0103 04:11:57.543033 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerDied","Data":"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7"} Jan 03 04:11:59 crc kubenswrapper[4921]: I0103 04:11:59.557192 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerStarted","Data":"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca"} Jan 03 04:11:59 crc kubenswrapper[4921]: I0103 04:11:59.558628 4921 generic.go:334] "Generic (PLEG): container finished" podID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerID="23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076" exitCode=0 Jan 03 04:11:59 crc kubenswrapper[4921]: I0103 04:11:59.558665 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerDied","Data":"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076"} Jan 03 04:11:59 crc kubenswrapper[4921]: I0103 04:11:59.592836 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j772m" podStartSLOduration=1.879043202 podStartE2EDuration="6.59282188s" podCreationTimestamp="2026-01-03 04:11:53 +0000 UTC" firstStartedPulling="2026-01-03 04:11:54.514935957 +0000 UTC m=+1850.126362781" lastFinishedPulling="2026-01-03 04:11:59.228714635 +0000 UTC m=+1854.840141459" observedRunningTime="2026-01-03 04:11:59.59084899 +0000 UTC m=+1855.202275814" watchObservedRunningTime="2026-01-03 04:11:59.59282188 +0000 UTC m=+1855.204248704" Jan 03 04:12:00 crc kubenswrapper[4921]: I0103 04:12:00.568182 4921 generic.go:334] "Generic (PLEG): container finished" podID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerID="fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a" exitCode=0 Jan 03 04:12:00 crc kubenswrapper[4921]: I0103 04:12:00.568247 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerDied","Data":"fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a"} Jan 03 04:12:01 crc kubenswrapper[4921]: I0103 04:12:01.579594 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerStarted","Data":"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5"} Jan 03 04:12:01 crc kubenswrapper[4921]: I0103 04:12:01.600639 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s2w2g" podStartSLOduration=3.982553994 podStartE2EDuration="5.600621719s" podCreationTimestamp="2026-01-03 04:11:56 +0000 UTC" firstStartedPulling="2026-01-03 04:11:59.56003035 +0000 UTC m=+1855.171457174" lastFinishedPulling="2026-01-03 04:12:01.178098035 +0000 UTC m=+1856.789524899" observedRunningTime="2026-01-03 04:12:01.597713886 +0000 UTC m=+1857.209140710" watchObservedRunningTime="2026-01-03 04:12:01.600621719 +0000 UTC m=+1857.212048543" Jan 03 04:12:03 crc kubenswrapper[4921]: I0103 04:12:03.533846 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:03 crc kubenswrapper[4921]: I0103 04:12:03.533911 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:04 crc kubenswrapper[4921]: I0103 04:12:04.597558 4921 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j772m" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="registry-server" probeResult="failure" output=< Jan 03 04:12:04 crc kubenswrapper[4921]: timeout: failed to connect service ":50051" within 1s Jan 03 04:12:04 crc kubenswrapper[4921]: > Jan 03 04:12:04 crc kubenswrapper[4921]: I0103 04:12:04.890885 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:12:05 crc kubenswrapper[4921]: I0103 04:12:05.613204 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"a502de01a964a53dd63eef3d3e98682b6d337e7bc49194214129905dca85dd66"} Jan 03 04:12:06 crc kubenswrapper[4921]: I0103 04:12:06.707972 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:06 crc kubenswrapper[4921]: I0103 04:12:06.708245 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:06 crc kubenswrapper[4921]: I0103 04:12:06.761000 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:07 crc kubenswrapper[4921]: I0103 04:12:07.672394 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:07 crc kubenswrapper[4921]: I0103 04:12:07.714994 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.169029 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-rkjb5_b9839659-8190-4224-ba78-2ef853498135/kube-rbac-proxy/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.250619 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-rkjb5_b9839659-8190-4224-ba78-2ef853498135/controller/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.395537 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-5vrrb_b520d64a-8e06-4c1c-abf1-e55d308aaedd/frr-k8s-webhook-server/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.430417 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-frr-files/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.550871 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-frr-files/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.592234 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-reloader/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.597542 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-metrics/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.643636 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-reloader/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.778730 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-frr-files/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.779115 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-metrics/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.796384 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-reloader/0.log" Jan 03 04:12:08 crc kubenswrapper[4921]: I0103 04:12:08.824963 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-metrics/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.014252 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-frr-files/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.014785 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-metrics/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.035886 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/cp-reloader/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.054636 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/controller/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.183803 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/frr-metrics/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.207474 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/kube-rbac-proxy/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.218882 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/kube-rbac-proxy-frr/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.441088 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/reloader/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.480623 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-87945dc6b-c7r8v_6e371e5d-ca0b-4fb7-ae0f-02941a69d3f0/manager/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.550015 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmp6h_7614f71a-a040-40d4-9f0a-9ebee98a7aed/frr/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.640941 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s2w2g" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="registry-server" containerID="cri-o://34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5" gracePeriod=2 Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.679380 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5587bf6f4f-lb878_71acd931-ba27-4db9-85ef-cbe3a14d28b5/webhook-server/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.740305 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hbzjk_3b5f636d-3481-4563-a5a7-a5564c024ec5/kube-rbac-proxy/0.log" Jan 03 04:12:09 crc kubenswrapper[4921]: I0103 04:12:09.877017 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hbzjk_3b5f636d-3481-4563-a5a7-a5564c024ec5/speaker/0.log" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.048057 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.133354 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content\") pod \"f5939ae6-1c71-4953-831f-604ea8dccbef\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.133450 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj2g4\" (UniqueName: \"kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4\") pod \"f5939ae6-1c71-4953-831f-604ea8dccbef\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.133484 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities\") pod \"f5939ae6-1c71-4953-831f-604ea8dccbef\" (UID: \"f5939ae6-1c71-4953-831f-604ea8dccbef\") " Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.134533 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities" (OuterVolumeSpecName: "utilities") pod "f5939ae6-1c71-4953-831f-604ea8dccbef" (UID: "f5939ae6-1c71-4953-831f-604ea8dccbef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.140675 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4" (OuterVolumeSpecName: "kube-api-access-sj2g4") pod "f5939ae6-1c71-4953-831f-604ea8dccbef" (UID: "f5939ae6-1c71-4953-831f-604ea8dccbef"). InnerVolumeSpecName "kube-api-access-sj2g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.159665 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5939ae6-1c71-4953-831f-604ea8dccbef" (UID: "f5939ae6-1c71-4953-831f-604ea8dccbef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.235098 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.235145 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5939ae6-1c71-4953-831f-604ea8dccbef-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.235159 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj2g4\" (UniqueName: \"kubernetes.io/projected/f5939ae6-1c71-4953-831f-604ea8dccbef-kube-api-access-sj2g4\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.659970 4921 generic.go:334] "Generic (PLEG): container finished" podID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerID="34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5" exitCode=0 Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.660012 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerDied","Data":"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5"} Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.660043 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s2w2g" event={"ID":"f5939ae6-1c71-4953-831f-604ea8dccbef","Type":"ContainerDied","Data":"71531f6245d4868df60d01cb75e7779dba60fcf7a53ec32d9d275b58259f3f86"} Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.660081 4921 scope.go:117] "RemoveContainer" containerID="34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.660080 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s2w2g" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.689464 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.692869 4921 scope.go:117] "RemoveContainer" containerID="fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.697157 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s2w2g"] Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.719756 4921 scope.go:117] "RemoveContainer" containerID="23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.742151 4921 scope.go:117] "RemoveContainer" containerID="34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5" Jan 03 04:12:10 crc kubenswrapper[4921]: E0103 04:12:10.742942 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5\": container with ID starting with 34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5 not found: ID does not exist" containerID="34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.742979 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5"} err="failed to get container status \"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5\": rpc error: code = NotFound desc = could not find container \"34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5\": container with ID starting with 34fe0d6ba4859eeb2fc915c57e82611a321d35cec74382232b156cfa0b0fe6b5 not found: ID does not exist" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.743026 4921 scope.go:117] "RemoveContainer" containerID="fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a" Jan 03 04:12:10 crc kubenswrapper[4921]: E0103 04:12:10.743292 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a\": container with ID starting with fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a not found: ID does not exist" containerID="fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.743312 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a"} err="failed to get container status \"fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a\": rpc error: code = NotFound desc = could not find container \"fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a\": container with ID starting with fc568246a8fb80417d7bd567833c53bc8c10882287d0bbc66832192980c03f9a not found: ID does not exist" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.743323 4921 scope.go:117] "RemoveContainer" containerID="23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076" Jan 03 04:12:10 crc kubenswrapper[4921]: E0103 04:12:10.743797 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076\": container with ID starting with 23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076 not found: ID does not exist" containerID="23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.743852 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076"} err="failed to get container status \"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076\": rpc error: code = NotFound desc = could not find container \"23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076\": container with ID starting with 23d48c5b2996def6d477435e7a0616acf20d39d0ee9c75e540923b155e45d076 not found: ID does not exist" Jan 03 04:12:10 crc kubenswrapper[4921]: I0103 04:12:10.891832 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" path="/var/lib/kubelet/pods/f5939ae6-1c71-4953-831f-604ea8dccbef/volumes" Jan 03 04:12:13 crc kubenswrapper[4921]: I0103 04:12:13.595153 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:13 crc kubenswrapper[4921]: I0103 04:12:13.639599 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:13 crc kubenswrapper[4921]: I0103 04:12:13.834567 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:12:14 crc kubenswrapper[4921]: I0103 04:12:14.691462 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j772m" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="registry-server" containerID="cri-o://1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca" gracePeriod=2 Jan 03 04:12:15 crc kubenswrapper[4921]: E0103 04:12:15.125218 4921 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb8a2128_f4a3_4cda_9b90_d57d95df0e89.slice/crio-conmon-1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca.scope\": RecentStats: unable to find data in memory cache]" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.583634 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.699898 4921 generic.go:334] "Generic (PLEG): container finished" podID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerID="1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca" exitCode=0 Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.699948 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j772m" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.699969 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerDied","Data":"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca"} Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.700320 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j772m" event={"ID":"eb8a2128-f4a3-4cda-9b90-d57d95df0e89","Type":"ContainerDied","Data":"1338e8b9083016002a017b59970fa715fe9a7b1093a683f014d0ffe2f2a2936a"} Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.700337 4921 scope.go:117] "RemoveContainer" containerID="1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.710574 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content\") pod \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.710625 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrkdc\" (UniqueName: \"kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc\") pod \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.710682 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities\") pod \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\" (UID: \"eb8a2128-f4a3-4cda-9b90-d57d95df0e89\") " Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.711889 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities" (OuterVolumeSpecName: "utilities") pod "eb8a2128-f4a3-4cda-9b90-d57d95df0e89" (UID: "eb8a2128-f4a3-4cda-9b90-d57d95df0e89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.721489 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc" (OuterVolumeSpecName: "kube-api-access-hrkdc") pod "eb8a2128-f4a3-4cda-9b90-d57d95df0e89" (UID: "eb8a2128-f4a3-4cda-9b90-d57d95df0e89"). InnerVolumeSpecName "kube-api-access-hrkdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.725530 4921 scope.go:117] "RemoveContainer" containerID="7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.765305 4921 scope.go:117] "RemoveContainer" containerID="f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.807778 4921 scope.go:117] "RemoveContainer" containerID="1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca" Jan 03 04:12:15 crc kubenswrapper[4921]: E0103 04:12:15.808309 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca\": container with ID starting with 1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca not found: ID does not exist" containerID="1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.808373 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca"} err="failed to get container status \"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca\": rpc error: code = NotFound desc = could not find container \"1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca\": container with ID starting with 1bb03cf7c54fd5c1011892b900699ea016f109d359b042f3401f379b2d67eeca not found: ID does not exist" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.808405 4921 scope.go:117] "RemoveContainer" containerID="7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7" Jan 03 04:12:15 crc kubenswrapper[4921]: E0103 04:12:15.810169 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7\": container with ID starting with 7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7 not found: ID does not exist" containerID="7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.810216 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7"} err="failed to get container status \"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7\": rpc error: code = NotFound desc = could not find container \"7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7\": container with ID starting with 7bfe0c2520789b2ea87d466647fa080ea6fc397ffb8fde86a513d2e8067d82e7 not found: ID does not exist" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.810245 4921 scope.go:117] "RemoveContainer" containerID="f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0" Jan 03 04:12:15 crc kubenswrapper[4921]: E0103 04:12:15.810636 4921 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0\": container with ID starting with f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0 not found: ID does not exist" containerID="f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.810663 4921 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0"} err="failed to get container status \"f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0\": rpc error: code = NotFound desc = could not find container \"f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0\": container with ID starting with f1145e466e0c786665c5c99bdbe5050fffee60c49134f349b5b9b52a94599bd0 not found: ID does not exist" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.814020 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrkdc\" (UniqueName: \"kubernetes.io/projected/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-kube-api-access-hrkdc\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.814057 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.848538 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb8a2128-f4a3-4cda-9b90-d57d95df0e89" (UID: "eb8a2128-f4a3-4cda-9b90-d57d95df0e89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:12:15 crc kubenswrapper[4921]: I0103 04:12:15.915614 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb8a2128-f4a3-4cda-9b90-d57d95df0e89-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:12:16 crc kubenswrapper[4921]: I0103 04:12:16.030320 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:12:16 crc kubenswrapper[4921]: I0103 04:12:16.035328 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j772m"] Jan 03 04:12:16 crc kubenswrapper[4921]: I0103 04:12:16.890297 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" path="/var/lib/kubelet/pods/eb8a2128-f4a3-4cda-9b90-d57d95df0e89/volumes" Jan 03 04:12:23 crc kubenswrapper[4921]: I0103 04:12:23.968827 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-09aa-account-create-qmd7n_015d13a2-fc39-4b75-b4ec-8d5bf46c4c00/mariadb-account-create/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.229011 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-lw567_a783c4d2-f5e1-4b54-bd67-e99180296cbc/glance-db-sync/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.303704 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-wnxlc_b1054977-0f34-413d-958f-5cdfe6c881f7/mariadb-database-create/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.395912 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_8a534b6a-d5ee-479c-b105-de684ca6fa62/glance-api/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.455789 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_8a534b6a-d5ee-479c-b105-de684ca6fa62/glance-log/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.462842 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-external-api-0_8a534b6a-d5ee-479c-b105-de684ca6fa62/glance-httpd/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.583509 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d48a3402-3b34-4080-bc8f-f1ad53021216/glance-api/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.637393 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d48a3402-3b34-4080-bc8f-f1ad53021216/glance-httpd/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.643100 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-internal-api-0_d48a3402-3b34-4080-bc8f-f1ad53021216/glance-log/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.853145 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-cron-29456881-9mwqb_805af972-9220-4b06-a94a-bca97464b177/keystone-cron/0.log" Jan 03 04:12:24 crc kubenswrapper[4921]: I0103 04:12:24.966377 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-79d9b68446-kxlwk_2a0b1317-5c69-485b-af20-2c12a9eb1743/keystone-api/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.162764 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_615f93ea-16ce-45da-88e6-237f1d57716a/mysql-bootstrap/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.344992 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_615f93ea-16ce-45da-88e6-237f1d57716a/mysql-bootstrap/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.433963 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_615f93ea-16ce-45da-88e6-237f1d57716a/galera/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.547758 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_2a9d550e-62e6-4dee-a034-d3e1ea269820/mysql-bootstrap/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.700302 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_2a9d550e-62e6-4dee-a034-d3e1ea269820/mysql-bootstrap/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.744277 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_2a9d550e-62e6-4dee-a034-d3e1ea269820/galera/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.894746 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_5dac2bf9-ef51-4e90-adab-2a73b463967a/mysql-bootstrap/0.log" Jan 03 04:12:25 crc kubenswrapper[4921]: I0103 04:12:25.926842 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_67fd2f86-99b0-4dfa-a40e-f3438940fa71/memcached/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.110803 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_5dac2bf9-ef51-4e90-adab-2a73b463967a/mysql-bootstrap/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.145084 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_63d57045-4c96-4700-8450-8c3bdf348bd3/openstackclient/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.168208 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_5dac2bf9-ef51-4e90-adab-2a73b463967a/galera/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.363168 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_e3b45f46-86ca-4ffb-b451-00d8631b9aaa/setup-container/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.555944 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_e3b45f46-86ca-4ffb-b451-00d8631b9aaa/setup-container/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.598030 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_e3b45f46-86ca-4ffb-b451-00d8631b9aaa/rabbitmq/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.603073 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-5957d6665c-x9qr6_61567511-53a7-47d2-8b71-e57910f99a18/proxy-httpd/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.791715 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-5957d6665c-x9qr6_61567511-53a7-47d2-8b71-e57910f99a18/proxy-server/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.833919 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-bckss_37fd922c-8a9f-4e29-8af5-54e73065236b/swift-ring-rebalance/0.log" Jan 03 04:12:26 crc kubenswrapper[4921]: I0103 04:12:26.995617 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/account-auditor/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.065468 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/account-reaper/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.069690 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/account-replicator/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.098841 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/account-server/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.190373 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/container-auditor/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.250978 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/container-server/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.270841 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/container-replicator/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.281935 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/container-updater/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.354099 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/object-auditor/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.401785 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/object-expirer/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.461996 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/object-replicator/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.520041 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/object-server/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.546473 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/object-updater/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.582716 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/rsync/0.log" Jan 03 04:12:27 crc kubenswrapper[4921]: I0103 04:12:27.643813 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_5b2d14db-f843-421e-895a-92dc3986c2c7/swift-recon-cron/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.413140 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/util/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.621685 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/pull/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.640182 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/pull/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.650249 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/util/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.790368 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/util/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.814861 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/extract/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.821289 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4wwlt4_da345cb0-e3ff-4397-bc44-1e0e4b419c59/pull/0.log" Jan 03 04:12:41 crc kubenswrapper[4921]: I0103 04:12:41.961929 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.090229 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.109877 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-content/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.110149 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-content/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.300716 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-content/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.301915 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.544577 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.710608 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kxb2v_c05c8aa8-a90c-45ec-9d86-72760b49dd24/registry-server/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.754653 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-content/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.772905 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.795611 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-content/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.917384 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-utilities/0.log" Jan 03 04:12:42 crc kubenswrapper[4921]: I0103 04:12:42.936808 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/extract-content/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.163667 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8s84l_549a7ea1-10f1-4e90-a2e3-9ef968ce88da/marketplace-operator/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.261004 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-utilities/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.309146 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bmtsf_0c4a9ea4-8edd-42c3-af24-d96bb6d8e4e4/registry-server/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.430404 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-utilities/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.435647 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-content/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.472137 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-content/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.607411 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-utilities/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.627358 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/extract-content/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.738893 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kkq2b_107e4592-8633-46d9-867c-c56028d7b8e8/registry-server/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.811693 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-utilities/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.976199 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-content/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.984460 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-utilities/0.log" Jan 03 04:12:43 crc kubenswrapper[4921]: I0103 04:12:43.997947 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-content/0.log" Jan 03 04:12:44 crc kubenswrapper[4921]: I0103 04:12:44.105613 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-utilities/0.log" Jan 03 04:12:44 crc kubenswrapper[4921]: I0103 04:12:44.168174 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/extract-content/0.log" Jan 03 04:12:44 crc kubenswrapper[4921]: I0103 04:12:44.607149 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-r5dqt_e0c419ea-fdfa-4b5a-90ff-9bcbeb5a0286/registry-server/0.log" Jan 03 04:12:49 crc kubenswrapper[4921]: I0103 04:12:49.043196 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-wnxlc"] Jan 03 04:12:49 crc kubenswrapper[4921]: I0103 04:12:49.049231 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-wnxlc"] Jan 03 04:12:50 crc kubenswrapper[4921]: I0103 04:12:50.891478 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1054977-0f34-413d-958f-5cdfe6c881f7" path="/var/lib/kubelet/pods/b1054977-0f34-413d-958f-5cdfe6c881f7/volumes" Jan 03 04:12:59 crc kubenswrapper[4921]: I0103 04:12:59.029540 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-09aa-account-create-qmd7n"] Jan 03 04:12:59 crc kubenswrapper[4921]: I0103 04:12:59.038091 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-09aa-account-create-qmd7n"] Jan 03 04:13:00 crc kubenswrapper[4921]: I0103 04:13:00.897720 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="015d13a2-fc39-4b75-b4ec-8d5bf46c4c00" path="/var/lib/kubelet/pods/015d13a2-fc39-4b75-b4ec-8d5bf46c4c00/volumes" Jan 03 04:13:07 crc kubenswrapper[4921]: I0103 04:13:07.044624 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lw567"] Jan 03 04:13:07 crc kubenswrapper[4921]: I0103 04:13:07.052604 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-lw567"] Jan 03 04:13:08 crc kubenswrapper[4921]: I0103 04:13:08.898600 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a783c4d2-f5e1-4b54-bd67-e99180296cbc" path="/var/lib/kubelet/pods/a783c4d2-f5e1-4b54-bd67-e99180296cbc/volumes" Jan 03 04:13:17 crc kubenswrapper[4921]: I0103 04:13:17.192189 4921 scope.go:117] "RemoveContainer" containerID="49202dc7df1ec6ce9150f547f9fc05bd234450bb71407f0429d6183ccc65412e" Jan 03 04:13:17 crc kubenswrapper[4921]: I0103 04:13:17.251863 4921 scope.go:117] "RemoveContainer" containerID="34411416145a1f7bb1308a53deffd0d5c974995fadccc0f1cbd653f54578633e" Jan 03 04:13:17 crc kubenswrapper[4921]: I0103 04:13:17.326346 4921 scope.go:117] "RemoveContainer" containerID="3553b26bf3b71df953c4ffe305d0272e0c5ff024c72e315af8b2a89535517cc2" Jan 03 04:13:51 crc kubenswrapper[4921]: I0103 04:13:51.554008 4921 generic.go:334] "Generic (PLEG): container finished" podID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerID="09e7bda62f70b51fe5c2be927c44df2555fa7f17a9e2d62b2affb323facd5ad4" exitCode=0 Jan 03 04:13:51 crc kubenswrapper[4921]: I0103 04:13:51.554089 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-g8xn7/must-gather-x86lq" event={"ID":"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9","Type":"ContainerDied","Data":"09e7bda62f70b51fe5c2be927c44df2555fa7f17a9e2d62b2affb323facd5ad4"} Jan 03 04:13:51 crc kubenswrapper[4921]: I0103 04:13:51.557662 4921 scope.go:117] "RemoveContainer" containerID="09e7bda62f70b51fe5c2be927c44df2555fa7f17a9e2d62b2affb323facd5ad4" Jan 03 04:13:52 crc kubenswrapper[4921]: I0103 04:13:52.148663 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g8xn7_must-gather-x86lq_ec47b9ac-0b5b-40ab-a738-d81d4425b2f9/gather/0.log" Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.465776 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-g8xn7/must-gather-x86lq"] Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.466774 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-g8xn7/must-gather-x86lq" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="copy" containerID="cri-o://29c5b4937cc4fa0759a4eb153bbf02fa2e9fa52a6dfd51db928ef82bac27fd3b" gracePeriod=2 Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.473746 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-g8xn7/must-gather-x86lq"] Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.647048 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g8xn7_must-gather-x86lq_ec47b9ac-0b5b-40ab-a738-d81d4425b2f9/copy/0.log" Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.647796 4921 generic.go:334] "Generic (PLEG): container finished" podID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerID="29c5b4937cc4fa0759a4eb153bbf02fa2e9fa52a6dfd51db928ef82bac27fd3b" exitCode=143 Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.847502 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g8xn7_must-gather-x86lq_ec47b9ac-0b5b-40ab-a738-d81d4425b2f9/copy/0.log" Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.849005 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.972035 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzhvh\" (UniqueName: \"kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh\") pod \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.972089 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output\") pod \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\" (UID: \"ec47b9ac-0b5b-40ab-a738-d81d4425b2f9\") " Jan 03 04:13:59 crc kubenswrapper[4921]: I0103 04:13:59.985803 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh" (OuterVolumeSpecName: "kube-api-access-fzhvh") pod "ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" (UID: "ec47b9ac-0b5b-40ab-a738-d81d4425b2f9"). InnerVolumeSpecName "kube-api-access-fzhvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.067786 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" (UID: "ec47b9ac-0b5b-40ab-a738-d81d4425b2f9"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.074063 4921 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.074090 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzhvh\" (UniqueName: \"kubernetes.io/projected/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9-kube-api-access-fzhvh\") on node \"crc\" DevicePath \"\"" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.661777 4921 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-g8xn7_must-gather-x86lq_ec47b9ac-0b5b-40ab-a738-d81d4425b2f9/copy/0.log" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.663361 4921 scope.go:117] "RemoveContainer" containerID="29c5b4937cc4fa0759a4eb153bbf02fa2e9fa52a6dfd51db928ef82bac27fd3b" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.663426 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-g8xn7/must-gather-x86lq" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.690113 4921 scope.go:117] "RemoveContainer" containerID="09e7bda62f70b51fe5c2be927c44df2555fa7f17a9e2d62b2affb323facd5ad4" Jan 03 04:14:00 crc kubenswrapper[4921]: I0103 04:14:00.892821 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" path="/var/lib/kubelet/pods/ec47b9ac-0b5b-40ab-a738-d81d4425b2f9/volumes" Jan 03 04:14:31 crc kubenswrapper[4921]: I0103 04:14:31.175664 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:14:31 crc kubenswrapper[4921]: I0103 04:14:31.176222 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.693386 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694201 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="extract-utilities" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694216 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="extract-utilities" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694226 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="gather" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694232 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="gather" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694251 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="extract-content" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694258 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="extract-content" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694292 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="extract-utilities" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694304 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="extract-utilities" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694315 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694323 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694331 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694338 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694362 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="copy" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694371 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="copy" Jan 03 04:14:53 crc kubenswrapper[4921]: E0103 04:14:53.694385 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="extract-content" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694392 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="extract-content" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694538 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb8a2128-f4a3-4cda-9b90-d57d95df0e89" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694578 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="copy" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694593 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5939ae6-1c71-4953-831f-604ea8dccbef" containerName="registry-server" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.694608 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec47b9ac-0b5b-40ab-a738-d81d4425b2f9" containerName="gather" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.696027 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.703189 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.752713 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghl82\" (UniqueName: \"kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.753023 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.753385 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.855281 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.855373 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.855399 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghl82\" (UniqueName: \"kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.856037 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.856109 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:53 crc kubenswrapper[4921]: I0103 04:14:53.875056 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghl82\" (UniqueName: \"kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82\") pod \"certified-operators-mkzz2\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:54 crc kubenswrapper[4921]: I0103 04:14:54.023448 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:14:54 crc kubenswrapper[4921]: I0103 04:14:54.496092 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:14:55 crc kubenswrapper[4921]: I0103 04:14:55.168477 4921 generic.go:334] "Generic (PLEG): container finished" podID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerID="9fb52448b6300e1be76fd3dcf2d791321927d6aa41c6a4a93766ed103aa6d88d" exitCode=0 Jan 03 04:14:55 crc kubenswrapper[4921]: I0103 04:14:55.168664 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerDied","Data":"9fb52448b6300e1be76fd3dcf2d791321927d6aa41c6a4a93766ed103aa6d88d"} Jan 03 04:14:55 crc kubenswrapper[4921]: I0103 04:14:55.168742 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerStarted","Data":"1197b718ee7b84cf7bf9aa3584554fdf6426dff0e0f536f1dd9293da296249b6"} Jan 03 04:14:56 crc kubenswrapper[4921]: I0103 04:14:56.178282 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerStarted","Data":"57802bc8069625b6291b932d05ff1a7b7339bf52d52d9e059df58c2bd11d370d"} Jan 03 04:14:57 crc kubenswrapper[4921]: I0103 04:14:57.191586 4921 generic.go:334] "Generic (PLEG): container finished" podID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerID="57802bc8069625b6291b932d05ff1a7b7339bf52d52d9e059df58c2bd11d370d" exitCode=0 Jan 03 04:14:57 crc kubenswrapper[4921]: I0103 04:14:57.191670 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerDied","Data":"57802bc8069625b6291b932d05ff1a7b7339bf52d52d9e059df58c2bd11d370d"} Jan 03 04:14:58 crc kubenswrapper[4921]: I0103 04:14:58.205587 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerStarted","Data":"5b5d0b2a4c41b8caaf251aba6b002c684ec11159c9911abff354b83daeabce31"} Jan 03 04:14:58 crc kubenswrapper[4921]: I0103 04:14:58.246460 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mkzz2" podStartSLOduration=2.756015103 podStartE2EDuration="5.246428995s" podCreationTimestamp="2026-01-03 04:14:53 +0000 UTC" firstStartedPulling="2026-01-03 04:14:55.170600831 +0000 UTC m=+2030.782027655" lastFinishedPulling="2026-01-03 04:14:57.661014683 +0000 UTC m=+2033.272441547" observedRunningTime="2026-01-03 04:14:58.236909961 +0000 UTC m=+2033.848336845" watchObservedRunningTime="2026-01-03 04:14:58.246428995 +0000 UTC m=+2033.857855849" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.151550 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l"] Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.153057 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.155806 4921 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.156454 4921 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.168972 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l"] Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.254643 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.254702 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcdqk\" (UniqueName: \"kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.254725 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.356057 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.357526 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcdqk\" (UniqueName: \"kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.357578 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.358034 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.363334 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.379229 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcdqk\" (UniqueName: \"kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk\") pod \"collect-profiles-29456895-jzw5l\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.472487 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:00 crc kubenswrapper[4921]: I0103 04:15:00.980953 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l"] Jan 03 04:15:00 crc kubenswrapper[4921]: W0103 04:15:00.988716 4921 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c6d4d36_af54_4586_a2dd_0d3a8b1f3619.slice/crio-bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9 WatchSource:0}: Error finding container bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9: Status 404 returned error can't find the container with id bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9 Jan 03 04:15:01 crc kubenswrapper[4921]: I0103 04:15:01.175495 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:15:01 crc kubenswrapper[4921]: I0103 04:15:01.175908 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:15:01 crc kubenswrapper[4921]: I0103 04:15:01.228942 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" event={"ID":"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619","Type":"ContainerStarted","Data":"bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9"} Jan 03 04:15:02 crc kubenswrapper[4921]: I0103 04:15:02.240744 4921 generic.go:334] "Generic (PLEG): container finished" podID="2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" containerID="140688bbc82fd37636d4120b3e47724ef2daad76a40e3049dbf2e2b432535665" exitCode=0 Jan 03 04:15:02 crc kubenswrapper[4921]: I0103 04:15:02.240809 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" event={"ID":"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619","Type":"ContainerDied","Data":"140688bbc82fd37636d4120b3e47724ef2daad76a40e3049dbf2e2b432535665"} Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.560221 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.642257 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume\") pod \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.642386 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume\") pod \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.642517 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcdqk\" (UniqueName: \"kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk\") pod \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\" (UID: \"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619\") " Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.643489 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" (UID: "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.648991 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" (UID: "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.649847 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk" (OuterVolumeSpecName: "kube-api-access-tcdqk") pod "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" (UID: "2c6d4d36-af54-4586-a2dd-0d3a8b1f3619"). InnerVolumeSpecName "kube-api-access-tcdqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.744336 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcdqk\" (UniqueName: \"kubernetes.io/projected/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-kube-api-access-tcdqk\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.744374 4921 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-config-volume\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:03 crc kubenswrapper[4921]: I0103 04:15:03.744387 4921 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c6d4d36-af54-4586-a2dd-0d3a8b1f3619-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.023643 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.024318 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.084724 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.272822 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.272811 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29456895-jzw5l" event={"ID":"2c6d4d36-af54-4586-a2dd-0d3a8b1f3619","Type":"ContainerDied","Data":"bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9"} Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.272938 4921 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb7cb4b4fb3ae2f43136cecf5e7e6f7a39f663611236ebb796debde2d339dbc9" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.349502 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.397192 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.652662 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87"] Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.664606 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29456850-jlf87"] Jan 03 04:15:04 crc kubenswrapper[4921]: I0103 04:15:04.900594 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aad5cc0-feb3-45ed-b12d-a9cad9e3564c" path="/var/lib/kubelet/pods/4aad5cc0-feb3-45ed-b12d-a9cad9e3564c/volumes" Jan 03 04:15:06 crc kubenswrapper[4921]: I0103 04:15:06.296707 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mkzz2" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="registry-server" containerID="cri-o://5b5d0b2a4c41b8caaf251aba6b002c684ec11159c9911abff354b83daeabce31" gracePeriod=2 Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.326030 4921 generic.go:334] "Generic (PLEG): container finished" podID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerID="5b5d0b2a4c41b8caaf251aba6b002c684ec11159c9911abff354b83daeabce31" exitCode=0 Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.326125 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerDied","Data":"5b5d0b2a4c41b8caaf251aba6b002c684ec11159c9911abff354b83daeabce31"} Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.600589 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.733496 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghl82\" (UniqueName: \"kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82\") pod \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.733595 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities\") pod \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.733667 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content\") pod \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\" (UID: \"642eca8a-4ba5-4f38-b30b-c04308b3a96c\") " Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.738402 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities" (OuterVolumeSpecName: "utilities") pod "642eca8a-4ba5-4f38-b30b-c04308b3a96c" (UID: "642eca8a-4ba5-4f38-b30b-c04308b3a96c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.744567 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82" (OuterVolumeSpecName: "kube-api-access-ghl82") pod "642eca8a-4ba5-4f38-b30b-c04308b3a96c" (UID: "642eca8a-4ba5-4f38-b30b-c04308b3a96c"). InnerVolumeSpecName "kube-api-access-ghl82". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.816200 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "642eca8a-4ba5-4f38-b30b-c04308b3a96c" (UID: "642eca8a-4ba5-4f38-b30b-c04308b3a96c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.835903 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghl82\" (UniqueName: \"kubernetes.io/projected/642eca8a-4ba5-4f38-b30b-c04308b3a96c-kube-api-access-ghl82\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.835937 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:08 crc kubenswrapper[4921]: I0103 04:15:08.835946 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/642eca8a-4ba5-4f38-b30b-c04308b3a96c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.337190 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkzz2" event={"ID":"642eca8a-4ba5-4f38-b30b-c04308b3a96c","Type":"ContainerDied","Data":"1197b718ee7b84cf7bf9aa3584554fdf6426dff0e0f536f1dd9293da296249b6"} Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.337315 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkzz2" Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.337324 4921 scope.go:117] "RemoveContainer" containerID="5b5d0b2a4c41b8caaf251aba6b002c684ec11159c9911abff354b83daeabce31" Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.377968 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.378879 4921 scope.go:117] "RemoveContainer" containerID="57802bc8069625b6291b932d05ff1a7b7339bf52d52d9e059df58c2bd11d370d" Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.392436 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mkzz2"] Jan 03 04:15:09 crc kubenswrapper[4921]: I0103 04:15:09.408734 4921 scope.go:117] "RemoveContainer" containerID="9fb52448b6300e1be76fd3dcf2d791321927d6aa41c6a4a93766ed103aa6d88d" Jan 03 04:15:10 crc kubenswrapper[4921]: I0103 04:15:10.898962 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" path="/var/lib/kubelet/pods/642eca8a-4ba5-4f38-b30b-c04308b3a96c/volumes" Jan 03 04:15:17 crc kubenswrapper[4921]: I0103 04:15:17.462877 4921 scope.go:117] "RemoveContainer" containerID="b2e39a43b31941ad6b4ac11ce08f580170f306366a0238d925cd36ec0a92f373" Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.176435 4921 patch_prober.go:28] interesting pod/machine-config-daemon-cctxw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.176936 4921 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.176987 4921 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.177816 4921 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a502de01a964a53dd63eef3d3e98682b6d337e7bc49194214129905dca85dd66"} pod="openshift-machine-config-operator/machine-config-daemon-cctxw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.177883 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" podUID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerName="machine-config-daemon" containerID="cri-o://a502de01a964a53dd63eef3d3e98682b6d337e7bc49194214129905dca85dd66" gracePeriod=600 Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.528904 4921 generic.go:334] "Generic (PLEG): container finished" podID="429ab47e-68f8-4b60-aa4c-ab79a764b7db" containerID="a502de01a964a53dd63eef3d3e98682b6d337e7bc49194214129905dca85dd66" exitCode=0 Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.529056 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerDied","Data":"a502de01a964a53dd63eef3d3e98682b6d337e7bc49194214129905dca85dd66"} Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.529748 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-cctxw" event={"ID":"429ab47e-68f8-4b60-aa4c-ab79a764b7db","Type":"ContainerStarted","Data":"ebcd548b3bb0b4afa9845b870696f3cfa87d79db39eee9e23ff5fdbd03c7fdd0"} Jan 03 04:15:31 crc kubenswrapper[4921]: I0103 04:15:31.529870 4921 scope.go:117] "RemoveContainer" containerID="542be617a76ff88d418d3f748ada374939e667bd3cafb1e38b43ea2690e1125f" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.436025 4921 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:15:45 crc kubenswrapper[4921]: E0103 04:15:45.436930 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="extract-utilities" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.436945 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="extract-utilities" Jan 03 04:15:45 crc kubenswrapper[4921]: E0103 04:15:45.436960 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" containerName="collect-profiles" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.436968 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" containerName="collect-profiles" Jan 03 04:15:45 crc kubenswrapper[4921]: E0103 04:15:45.436984 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="registry-server" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.436994 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="registry-server" Jan 03 04:15:45 crc kubenswrapper[4921]: E0103 04:15:45.437009 4921 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="extract-content" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.437016 4921 state_mem.go:107] "Deleted CPUSet assignment" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="extract-content" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.437179 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c6d4d36-af54-4586-a2dd-0d3a8b1f3619" containerName="collect-profiles" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.437202 4921 memory_manager.go:354] "RemoveStaleState removing state" podUID="642eca8a-4ba5-4f38-b30b-c04308b3a96c" containerName="registry-server" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.438623 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.481357 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.546865 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crxfd\" (UniqueName: \"kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.546951 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.546998 4921 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.649559 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crxfd\" (UniqueName: \"kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.649641 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.649679 4921 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.650185 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.650445 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.678641 4921 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crxfd\" (UniqueName: \"kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd\") pod \"community-operators-trcs4\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:45 crc kubenswrapper[4921]: I0103 04:15:45.783526 4921 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:46 crc kubenswrapper[4921]: I0103 04:15:46.100949 4921 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:15:46 crc kubenswrapper[4921]: I0103 04:15:46.654718 4921 generic.go:334] "Generic (PLEG): container finished" podID="f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" containerID="3e2b6ba24a2bd2d26175f382985036269a7a452a92d04634c758aaf987ecc3a0" exitCode=0 Jan 03 04:15:46 crc kubenswrapper[4921]: I0103 04:15:46.654761 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerDied","Data":"3e2b6ba24a2bd2d26175f382985036269a7a452a92d04634c758aaf987ecc3a0"} Jan 03 04:15:46 crc kubenswrapper[4921]: I0103 04:15:46.654791 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerStarted","Data":"6715bcfa6e3b99142626501dba394941f552a05d9d333a900d4b07636c919891"} Jan 03 04:15:48 crc kubenswrapper[4921]: I0103 04:15:48.671207 4921 generic.go:334] "Generic (PLEG): container finished" podID="f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" containerID="0860764b1f308700648825f385f3833711506e9f7c7da5e06318247e85f92834" exitCode=0 Jan 03 04:15:48 crc kubenswrapper[4921]: I0103 04:15:48.671336 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerDied","Data":"0860764b1f308700648825f385f3833711506e9f7c7da5e06318247e85f92834"} Jan 03 04:15:48 crc kubenswrapper[4921]: I0103 04:15:48.673106 4921 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 03 04:15:49 crc kubenswrapper[4921]: I0103 04:15:49.681759 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerStarted","Data":"cd0702dd24350f32826ebcbfa4e8f9fc00118a8642562013c9f56fa7cb20b468"} Jan 03 04:15:49 crc kubenswrapper[4921]: I0103 04:15:49.715741 4921 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-trcs4" podStartSLOduration=2.241167742 podStartE2EDuration="4.715716369s" podCreationTimestamp="2026-01-03 04:15:45 +0000 UTC" firstStartedPulling="2026-01-03 04:15:46.656331125 +0000 UTC m=+2082.267757949" lastFinishedPulling="2026-01-03 04:15:49.130879752 +0000 UTC m=+2084.742306576" observedRunningTime="2026-01-03 04:15:49.706827161 +0000 UTC m=+2085.318254045" watchObservedRunningTime="2026-01-03 04:15:49.715716369 +0000 UTC m=+2085.327143223" Jan 03 04:15:55 crc kubenswrapper[4921]: I0103 04:15:55.784604 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:55 crc kubenswrapper[4921]: I0103 04:15:55.786479 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:55 crc kubenswrapper[4921]: I0103 04:15:55.864649 4921 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:56 crc kubenswrapper[4921]: I0103 04:15:56.795783 4921 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:15:59 crc kubenswrapper[4921]: I0103 04:15:59.442667 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:15:59 crc kubenswrapper[4921]: I0103 04:15:59.444397 4921 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-trcs4" podUID="f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" containerName="registry-server" containerID="cri-o://cd0702dd24350f32826ebcbfa4e8f9fc00118a8642562013c9f56fa7cb20b468" gracePeriod=2 Jan 03 04:16:00 crc kubenswrapper[4921]: I0103 04:16:00.768250 4921 generic.go:334] "Generic (PLEG): container finished" podID="f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" containerID="cd0702dd24350f32826ebcbfa4e8f9fc00118a8642562013c9f56fa7cb20b468" exitCode=0 Jan 03 04:16:00 crc kubenswrapper[4921]: I0103 04:16:00.769503 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerDied","Data":"cd0702dd24350f32826ebcbfa4e8f9fc00118a8642562013c9f56fa7cb20b468"} Jan 03 04:16:00 crc kubenswrapper[4921]: I0103 04:16:00.955802 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.022950 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content\") pod \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.023026 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crxfd\" (UniqueName: \"kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd\") pod \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.023066 4921 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities\") pod \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\" (UID: \"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124\") " Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.024671 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities" (OuterVolumeSpecName: "utilities") pod "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" (UID: "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.031604 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd" (OuterVolumeSpecName: "kube-api-access-crxfd") pod "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" (UID: "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124"). InnerVolumeSpecName "kube-api-access-crxfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.076585 4921 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" (UID: "f8b5edd2-55f7-4f4d-852d-ebe5af3fd124"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.124683 4921 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.124737 4921 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crxfd\" (UniqueName: \"kubernetes.io/projected/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-kube-api-access-crxfd\") on node \"crc\" DevicePath \"\"" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.124753 4921 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124-utilities\") on node \"crc\" DevicePath \"\"" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.781877 4921 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-trcs4" event={"ID":"f8b5edd2-55f7-4f4d-852d-ebe5af3fd124","Type":"ContainerDied","Data":"6715bcfa6e3b99142626501dba394941f552a05d9d333a900d4b07636c919891"} Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.781940 4921 scope.go:117] "RemoveContainer" containerID="cd0702dd24350f32826ebcbfa4e8f9fc00118a8642562013c9f56fa7cb20b468" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.781994 4921 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-trcs4" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.815376 4921 scope.go:117] "RemoveContainer" containerID="0860764b1f308700648825f385f3833711506e9f7c7da5e06318247e85f92834" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.847965 4921 scope.go:117] "RemoveContainer" containerID="3e2b6ba24a2bd2d26175f382985036269a7a452a92d04634c758aaf987ecc3a0" Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.850884 4921 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:16:01 crc kubenswrapper[4921]: I0103 04:16:01.863991 4921 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-trcs4"] Jan 03 04:16:02 crc kubenswrapper[4921]: I0103 04:16:02.911110 4921 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8b5edd2-55f7-4f4d-852d-ebe5af3fd124" path="/var/lib/kubelet/pods/f8b5edd2-55f7-4f4d-852d-ebe5af3fd124/volumes" Jan 03 04:16:03 crc kubenswrapper[4921]: I0103 04:16:03.434967 4921 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/swift-proxy-5957d6665c-x9qr6" podUID="61567511-53a7-47d2-8b71-e57910f99a18" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515126114067024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015126114070017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015126107461016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015126107461015460 5ustar corecore